var/home/core/zuul-output/0000755000175000017500000000000015134632744014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134636547015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000231736115134636473020277 0ustar corecore;=sikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD F"mv?_eGbuuțx{w7ݭ7֫C% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{3C_3& TQ0Z%bb oHIl.f/M1FJdl!و4Gf#C2lIw]BPIjfkAubTI *JB4?PxQs# `LK3@g(C U {oLtiGgz֝$,z'vǛVB} eRB0R딏]dP>Li.`|!>ڌj+ACl21E^#QDuxGvZ4c$)9ӋrYWoxCNQWs]8M%3KpNGIrND}2SRCK.(^$0^@hH9%!40Jm>*Kdg?y7|&#)3+o,2s%R>!%*XC7Ln* wCƕH#FLzsѹ Xߛk׹1{,wŻ4v+(n^RϚOGO;5p Cj·1z_j( ,"z-Ee}t(QCuˠMkmi+2z5iݸ6C~z+_Ex$\}*9h>t m2m`QɢJ[a|$ᑨj:D+ʎ; 9Gacm_jY-y`)͐o΁GWo(C U ?}aK+d&?>Y;ufʕ"uZ0EyT0: =XVy#iEW&q]#v0nFNV-9JrdK\D2s&[#bE(mV9ىN囋{V5e1߯F1>9r;:J_T{*T\hVQxi0LZD T{ /WHc&)_`i=į`PÝr JovJw`纪}PSSii4wT (Dnm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁weor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/P_]F@?qr7@sON_}ۿ릶ytoyמseQv^sP3.sP1'Ns}d_ս=f1Jid % Jwe`40^|ǜd]z dJR-Дxq4lZ,Z[|e 'Ƙ$b2JOh k[b>¾h[;:>OM=y)֖[Sm5*_?$cjf `~ߛUIOvl/.4`P{d056 %w ^?sʫ"nK)D}O >%9r}1j#e[tRQ9*ء !ǨLJ- upƜ/4cY\[|Xs;ɾ7-<S1wg y &SL9qk;NP> ,wդjtah-j:_[;4Wg_0K>є0vNۈ/ze={< 1;/STcD,ڙ`[3XPo0TXx ZYޏ=S-ܑ2ƹڞ7կZ8m1`qAewQT*:ÊxtŨ!u}$K6tem@t):êtx: `)L`m GƂ%k1羨(zv:U!2`cV, lNdV5m$/KFS#0gLwNO6¨h}'XvوPkWn}/7d*1q* c0.$\+XND]P*84[߷Q뽃J޸8iD WPC4>9 *#LC ءzCwS%'m'3ܚ|otoʉ!9:PZ"ρ5M^kVځIX%G^{;+Fi7Z(ZN~;MM/u2}ݼPݫedKAd#[ BeMP6" YǨ 0vyv?7R F"}8&q]ows!Z!C4g*8n]rMQ ;N>Sr??Ӽ]\+hSQזL +wfm#Y~!%rpWMEWMjbn(ek~iQ)à/2,?O 'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44JٵK+Woc.F3 %N%FF"HH"\$ۤ_5UWd̡bh塘ZRI&{3TUFp/:4TƳ5[۲yzz+ 4D.Ճ`!TnPFp':.4dMFN=/5ܙz,4kA<:z7y0^} "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@DEu&ݛȘPˬ-Ő\B`xr`"F'Iٺ*DnA)yzr^!3Ír!S$,.:+d̋BʺJ#SX*8ҁW7~>oOFe-<uJQ|FZEP__gi(`0/ƍcv7go2G$ N%v$^^&Q 4AMbvvɀ1J{ڔhэK'9*W )IYO;E4z⛢79"hK{BFEmBAΛ3>IO j u߿d{=t-n3Pnef9[}=%G*9sX,¬xS&9'E&"/"ncx}"mV5tŘ:wcZ К G)]$mbXE ^ǽ8%>,0FЕ 6vAVKVCjrD25#Lrv?33Iam:xy`|Q'eű^\ơ' .gygSAixپ im41;P^azl5|JE2z=.wcMԧ ax& =`|#HQ*lS<.U׻`>ajϿ '!9MHK:9#s,jV剤C:LIeHJ"M8P,$N;a-zݸJWc :.<sR6 լ$gu4M*B(A ݖΑِ %H;S*ڳJt>$M!^*n3qESfU, Iĭb#UFJPvBgZvn aE5}~2E|=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞t1Rx}fvvPXdQSg6EDT:dׁz^DjXp͇G|X5Q9K$)U?o': .,wؓaՁ_ 3]Q16ZYafuvrq^ѷQT},!H]6{Jw>%wK{)rH+"B4H7-]r}7v8|׾~Us?yWfv3>xpRҧH-EeJ~4YIozi:nq Vq8swHOzf ̙eX-4`TDGq G.tݻgq74ŠqBFf8 9Fk Afq#ϛa$!qNCJ4bnvB @W,v&- 6wCBjxk9ᤉ ,Asy3YޜZ4ΓVYf'h?kNg?҆8oC!IMo:^G10EY↘H:L@D+dˠUHs[hiҕ|֏G/G`' m5p|:9U8PZ7Yݷ/7cs=v{lLHqyXR iE^1x5/[O6rpP40ޢE_A͝ Z5 om2p)lbp/bj_d{R\' 礅_}=\:Nb{}IStgq$<$ilb)n&  $uT{wD]2cM(%YjDktByxVl巳1~jpd1O9Á%˧Byd}gs9QNʟ. /ӦxbHHAni5(~p>/O0vEWZ nY3 cU $O,iLacoW1/W=-kqb>&IL6i}^^XpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋZ5NhfӛŮ Qu8Y4?W֫/&W˸~%pqq{% ?K~,#/0'NZ׽Kq^ėSJ6#j8GO[ PCbʍN^XS&}E9OZ]'t$=tnn&nu [}Ab4 +OLuU{0fIb { Ov8FHӜ"D$aǽO8'1lfYuB!!=?8[Y|-ɬeǪzd;-s~CM>e:9[_v~\:P ؇'k01Q1jlX)/ΏL+NhBUx~Ga>Z"Q_wjTLRˀtL L+BT҂ll魳cf[L̎`;rK+S- (J[(6 b F? ZvƂcW+dˍ-m𢛲@ms~}3ɱ© R$ T5%:zZ甎܋)`ŰJ38!;NfHohVbK :S50exU}W`upHЍE_fNTU*q%bq@/5q0);F74~'*z[\M-~#aSmMÉB2Nnʇ)bAg`u2t"8U [tJYSk, "vu\h1Yhl~[mhm+F(g 6+YtHgd/}7m]Q!Mę5bR!JbV>&w6οH+NL$]p>8UU>Ѫg39Yg>OF9V?SAT~:gGt $*}aQ.Zi~%K\rfm$%ɪq(%W>*Hg>KStE)KS1z2"h%^NEN?  hxnd/)O{,:خcX1nIaJ/t4J\bƀWc-d4M^d/ ʂK0`v%"s#PCoT/*,:[4b=]N&, ,B82^WK9EHLPm))2.9ȱ  QAcBC-|$M\^B!`}M^t+C~Lb }D>{N{Vt)tpDN,FCz~$)*417l;V iэ(_,j]$9O+/Sh]ice wy\Mڗ$,DJ|lj*à␻,?XAe0bX@ h0[}BU0v']#Vo !ې: Z%ƶ(fl>'"Bg< 0^_d0Y@2!ӸfZ{Ibi/^cygwדzY'Ź$:fr;)ٔf ՠ3Kcxwg*EQU{$Sڸ3x~ 5clgSAW"X Pҿ.ظwyV}̒KX9U1>V..W%GX +Uvzg=npu{do#Vb4ra\sNC/T"*!k愨}plm@+@gSUX覽t01:)6kSL9Ug6rEr(3{ xRP8_S( $?uk| ]bP\vۗ晋cgLz2r~MMp!~~h?ljUc>rw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q?A-?T_qOq?5-3 |q|w.dަ'/Y?> (<2y. ">8YAC| w&5fɹ(ȊVã50z)la.~LlQx[b&Pĥx BjIKn"@+z'}ũrDks^F\`%Di5~cZ*sXLqQ$q6v+jRcepO}[ s\VF5vROq%mX-RÈlб 6jf/AfN vRPػ.6<'"6dv .z{I>|&ׇ4Ăw4 [P{]"}r1殲)ߚA 2J1SGpw>ٕQѱ vb;pV ^WO+į1tq61W vzZ U'=҅}rZ:T#\_:ď);KX!LHuQ (6c94Ce|u$4a?"1] `Wa+m𢛲`Rs _I@U8jxɕͽf3[Pg%,IR Ř`QbmүcH&CLlvLҼé1ivGgJ+u7Τ!ljK1SpHR>:YF2cU(77eGG\ m#Tvmە8[,)4\\=V~?C~>_) cxF;;Ds'n [&8NJP5H2Զj{RC>he:ա+e/.I0\lWoӊĭYcxN^SPiMrFI_"*l§,̀+ å} .[c&SX( ( =X?D5ۙ@m cEpR?H0F>v6A*:W?*nzfw*B#d[se$U>tLNÔ+XX߇`cu0:U[tp^}{>H4z 4 (DtH-ʐ?sk7iIbΏ%T}v}e{aBs˞L=ilNeb]nltwfCEI"*S k`u ygz[~S [j3+sE.,uDΡ1R:Vݐ/CBc˾] shGՙf 2+);W{@dlG)%عF&4D&u.Im9c$A$Dfj-ء^6&#OȯTgرBӆI t[ 5)l>MR2ǂv JpU1cJpրj&*ߗEЍ0U#X) bpNVYSD1౱UR}UR,:lơ2<8"˓MlA2 KvP8 I7D Oj>;V|a|`U>D*KS;|:xI/ió21׭ȦS!e^t+28b$d:z4 .}gRcƈ^ʮC^0l[hl"য*6 ny!HQ=GOf"8vAq&*țTOWse~ (5TX%/8vS:w}[ą qf2Lυi lm/+QD4t.P*2V J`\g2%tJ4vX[7g"z{1|\*& >Vv:V^S7{{u%[^g=pn]Y#&ߓTί_z7e&ӃCx;xLh+NOEp";SB/eWٹ`64F 2AhF{Ɩ;>87DǍ-~e;\26Lة:*mUAN=VޮL> jwB}ѹ .MVfz0Ïd0l?7- }|>TT%9d-9UK=&l&~g&i"L{vrQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨Or4?Jƶ7_+Bs=8'kP 3 |}44S8UXi;f;VE7e4AdX-fS_Ui/L6$dts` P}KNf@r3 Zj/}I+ϟYSUkTa>x!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ {6_{TQu<:!,gUw+E=YR0tK12ƭSg?LD j%1Hc!KO6P&j$MQbJ-ig~jpva?T]x@CUBDQ,)"Dd׃?`EPeܚ:ۣ̊[yoO"CU 9x{A U)q8icA U=:n#}U)+ l=i c3 U] W*1<gb(|"/qKDQ_ǧG.x6"Cst 45p0C7x ;9' [ӵւתx)O]|658Ǧn:yn{3tó^o<(㰨-( ~":&2޴<>ۦkZe-B5'˵uwX}+{iqnn29o|'*tW^ro1d(lxIk~ /E&*%A5__gY,i9q5qc3,Wt1)yX#V,2Q@&a)y5MTb\fewm5,dQrڋRE+c͘c,GFhglݴ,q#6<3Q Xfg 4̳S :GDQ/[-џ"%jed!O&LiyY<ņ:>ŕu?#w 4mŚVʹDE!Q$r`W"z5L|UYV' >1B.$U31 (Q92b|GgBfU'wt7rR4k- ֺb/y:Y1ZdW-peh y=_RU%ZT_B~v/Y{j G$Ab-@L|U %n VԴ K vr v|MO¹b<:n~~ʋ ox R,yը%G?.G_ !n}ol##i4/)Adqr%n!XdЈR<)q*T_ֳ9wٞGm<Š#)ӴoGo0:l}Sp5VNq"q|D!n#!_"ΪrSP0˪ƕ$ 6#Hd 5KN\E[0:U!CXsZdhhBs-5w9iͶJD ީs O}OꤒEu4,jD?>Q8;mLJ'GIVUC*xdӒ1 >84 ?->[jZ1meU珟t+yI85$j4 õX_ۯ*NGnA%E c{T )f"8("1¶7ROQwnOE8l7A]es (CS]'1an dzU),YN0 l)O>9a~8)1Ώ#C<$(K>oĜ^aY<Ǐ,˕i{:l[bo75O` />;Yó:HEUMSMo{9=>}?~}Y~ܤH$1'JDG .iooM;6~^>C@< ʏ^l{WGT-I.OٌCQ]g{XW$0b2@m/' r a:Ǘ'rxBOn*'6ak.EgltSC[fg'h&‹~")XǸojƈv2uzx Յ{ b@Jvlo{%n =m>-egYT5O7pw؃D<_>m^GOZ#Az״"1 ,-D6}Gr]͈q׍K§%bp/w޿i"{$jԒ$V%6UB{@f#t:5Fߝ4چBSZdڗ1MX5uc%5-y76UOtWRҰD0MGT(k!hq!"E$ S+wi.Ņ9zvuhpdbeDjksY5BFS {*M7-itSÃ-4ıo-H-0Iǃj 0pB,#1SZCK^Ksh-UF WtWycTdzѵT-w-U? δj,KaU|E8.oՉtM./zI1Ҩz іSIUz/VWXfe_joXIt Li4^+ Ծʂ:26GlԖ5̶GI*MP\UV)LJQ؏U"9eu2KNۮEu[fLEh-`Z]}+sd.g뾀CslUźކc؄N8!#f A;*+qUϳmnHwU(7YrEx\pumZ[o}QPk`1% 7 R+LB@QpTꜧ2eEBbZ"-uT }("2WԴqIJc=C#jl0VA$ekMm(k4k7m[  [OA Ռ?RkƮiǗbڛ!hcM2uUgzSJT96\ol gn@KľxQd߹㿭fk9=ǫ3IVhMLxnmhV1CߜnG%5L,ke ?*kۣѭiKIǶbrM}mY+LqAv`XoAžs#2 㑧Os]y=oBV78WsUI-Z&FhvΕ J}ʾM` [,MV0T[)iS"(OkE g"}NWUo\{0T^rM.Fϵ7צliĘRǖӼf=oSbGO%eɢ)+ur*x޳C4}R[p}݌n?y2 ua/Z_{_e'0q aZbƫ _duC`諒=l/tUT- :IzHAp02g|S]n>A-k4ª;7c?_{؏̰  ,c=tw^4]աprōsQTy-G?:7yAfK5Q)ozeQCL]JZY3e*y$ҺT ęj8UgI6D3TٜHsw0ÖUZF^<?D6a-T  53U>wJn2&f1lWfL0ן6vFOq~N ę;p ^Q+>mݑ;ݙC:0n1k0ؙfx&~29ߝ\ITRvyƉ9OȠޭIXln&Xa̮\};p1s}c n̳ŢMmb̰wnc+s#k/c@?c=fve~/F=>f#bcËfvechg cao&2Sa8c ,cG{Ic/Hlhex|Jz=P02yLnb^z  ;s ooFѿY _7vn% }G[n?lD?ؓ#o;7\@y;3cuNg>v黬u#ڝb<ǏV v.}sFy:a<0vo wf֮B+tBN] s۶+PH9vzZ'vLP$eD]Rc3{R-;!myڿ%"{Č_j:NgRsvy'_ ;a]hJ5*S::}12MuQoTeklA(yo$+&ק2)D*^r*NN1y wFg cSOd\n*#?T2TD]:=WSű ֪U:5_85ep]_d9{zw6*"EUzoԪO_>QBOW?":~w"pݽ_0pB(AMc}pD ^,^(za';n;Ѕݯ;bs(bnػ$zlv?b^#* ^`z%zaޖ}MF0]߇0%@X"넂X^_'^ܯ8W"c=qSF,^/qI5@ < A 2~%Jྰ'b ]+Kvk2>5/(x?_i{i$ 7rz`~8k#;b&PzN!  <_!a3/݂xW}M%6a!r/X A}0!D!Xo.*쇗 @1Wn4bĀ~8c`*f qx G2 By//^wkQ~]i]˺5>]P9g-(+O0Ckrٮ5nw R<7n[+ lk4͔P>r9dݲwz5N9n` ܞ/9w{C8^(n8 /anr r>70'n3. ==tAOZAgDQo D"!(+3ɫMAV<[ORL/J؇Ax&яNX: 5}J{!/g5r=S=Cf1g=W )I1@;lvk0_'J9VF7)`HUC0)jR^)v9L ؙ$2 #gmYVR'&=TzR΅^ Kd+2+5x^'Ս ĈjWt5ĵ2\x2EI%2OrJeR 'hR矋w;SI`OFcmt T'rˤsit^MH{ jì` b5(qb3 ^ϋ7UءoSv8^91e?raՊ,O Jg3f3J(]tz>sVMhJqXVj ;wrE6Y<8S@,Fs j6pEvO @nl&RGW&Y=x/JMi0YQfg]̜PAgvxOg *l^U;֣&8@ s K&ܺ;KE0@4JnC)6ͩ._ ѺX,u<01 PȃϡuUG+&"70 Bw@#pأ/+8yfIyYUrӏ3WqgIw|@.eݠC3la_1wZń9Ʉ~ILO/NB\ $u;V}S/if4Z"@p˯h)tD?K;u~#7Il0/Y.lǵimڰuuEISؑzUǁ(Q;@8AhS7Q{^ɈtM^;QbPKh<ULɴOǁF`1k2]+i 趄'FՌ9!i'^n3pʻ oZfW>@sօ )ßmLML6Adچ07E֎ތ;iLwN.bYVVtp[NP⦆bq7b 0 *1S_Ʈ6( ^Mɟnq|"~;<|Z|ѷBgve՟*?O145HxWoxj ,rXHY4L93o27QI)9E`7 |ozv: k(p(7(t؜=JҀ Rf@Nsfʷa~u) {0Cߖ%?'zWy`KM#:}Џ6DjJw6+&h>  \&.t:`hrU,.傤%NV!t`1/Iy%] Ќ: LGZ@]J/l P;gO3>38ER}ݔ?9YH*Z̖*p">@2u{6%8 s}Hyp ~dT=]d~\a`hEBp^ խ Ȏ<4~\pO\3yT%X&hC-A Ac ũfrj<J]!"V7Fcx'#f4Mb+*ޗ?tAF 0CeiT4wXVlN땥Iz5A=)| j5 PqMZ=enQZ$˫@ %eCd?ܖF@xX~n@ `\8YN`UWq. eE}o[G旯׳qY9C y2_Zl9+%P:E>ͫdb xzwSi@5'le{%&P͸BQq\Hxk^.<-, X&"h bAo$yQ uX̎tSY+ =VқB`Ol4.XIXj.lo6^P z!S`!oދQN2 ~B6b14{5mKH/D8DYnAz'![P-LŢ2;*bZd4[qV07'"wͪoK3u[PY>.5.x6TIYϋEdAyxV"ˀ.2?gB㵻X3=ESȧVyT>TC9sxh̓44ʅUd SʄNV8^!V.|/EzmvUNWeumF"Pذ%Ar ] G袬FXcH0]Jȟ:O+ca[Si]q8VYy5J c P[I #&s3HZs]33H!Nou-nӥidkKHx(@c~3=|WkD|hCmǢ0vL'-6EDTEoxY&-4;حTumhY2~r;@C;|<F>AY=t 74&c㥸K POpT( qO\ٙm?KAsm;?4i :t1g$61AYdE:#5@DDZ(w嬸ޤn w_8}k ? m?ql >kXz'7㝔oNH~E2Q@Cm= a2a2 x), &3jQVq8Ӑ3CWJ SYGh עYi 4yJ0C!lEggQϮj6_Nм*m&jR-:,>[UIn+Sp9O 1Bi.4#g NW콓2_` /Tx;BPK*2Gѻ,s5])rNi5о,UrZ 4cZ>mi$*?0GK c9q7,y%|1鑻k{ttr,v֦"2DZdrc0+}tF wfDBUb>f"eYeLPͦYMAmnR^i\0KdG8Yy%)w|`„r]?jBGf]8ojpXtɺ).%VL.N *}VKX{ ZywzNQL毊i ]6f%2B漃 0bV,0zk'Wۘ4\3G,DsoM "e9G.N89aKLQ ]'9.@rnUUH4$zO|]IG䫺%2s<J1Ǝp 0(Kv(Ď2[ĿZ^R)L:]ז3m,/yWFʢ0`^}<%Z&i`FHZ챣(hȒU}̸ W}]W͏ӧٮ:Hs6p}W ǫ|)W%7^>I8?{K]E0~LOyƪ yFq?jojz.7cW>oϓtD;js핒 ^*z~-?57>_ۋgc-gϳpz^f^{o~:Dh. K& gxn 'uzݩ鏗S5Lc:bM'Ox_?`İn6[W颉wZmj2T>-v1DŽvΥĽSe4ӵzg*Zp`Q·{-Hb~ۯ,o h"蛵>ʬ 3,Y:岍']FR' 9V_64tCN{V惁h]3BI_0ؑ)!AЍE0cY_ ^>ۘ0U[DɶHz "!ۡ8g2bL&K1I|L6"F[=L#};P ~1-#!VDdW'Qv1 :_݁z褩icR$ĂwЧ5aH:;Ew$8GA)‹s %-8"4@0dF"!vqo\P-SEWOx3x8$J( O ÷gzϷc>b&ctJھ&F!4fbD^Izy_S88;hΤlW `B3OR!1F9t'1Q M3,x phVU9# sx"Q0Ni\ԮE5 Ɠm-qb`+QtंpshAmJbJ] yb$)hlOОѦ> o#M'=%}߅dܪX˝A'Fҙz$8Il!Ե2@?t_r?%JZI!5FyUT?_P(]T{ҩPǑgk2if=!b LZlN0I)f&gDN5Y%ӊ, u4!>iM61RL[ھkQ hhD g1?̢Gu 858!+yEL+*JeaUeyv[d%# BbL[^lc5tU ^QGVĈA{w/ RAAs;"Q X'P\LK^ბpcBd#CҁΎ(3#UB ޳MXհh!> *cW D_Hp IdH=c:5UXkP$7/7찮=ӌ1ܭ'4*֢6`2u =K`bc!0ˢHz1eUȜը@D9h)e۽Ì^SV^SXh=ҖәF)kuEcxByL(!ᙅʀ$xCh08ݢI)g¢ !CYaXF|캡<UYj) z۷WĦg?c. \u92`L bMKB*ˌ͵9 BL%ͺTvN`2}3:!;X/=J̴zQio>t_vb0ykT/kd׷ZB2B }_ y%IgO|[1! Y6CBL hץ"Icdx 48A~zykH`L2 hSw8JޘDJ:a#Yg/33mo^G7+YSfpə&tBo&37S9!3/\1cR0ƷφCS1 k@LaL/|HCh̉ ;FmpA@o), w\:Zrd_̕PgU مE8nYڔa [>N3 [480z]{W)P\Ёc0GOt-]/t <6UX%rDVXkKM.]6,k۾/b\E j0o/Ļ8}RZ;r/F/jpoVyz ?!Du,0QB +%Ĉx>;K!#>,gykHR9ŋP-),-k W 3s# [Ztp-X$e=фԩaI.!4{^ *wyeJTdVә*.4#|M$8Xt pfQDiY .NcD)W*EL =?ܬIp<1r%-uX,\Y?L{+qki\IȖqtxbH:l͚=#܉ǻqWw 4IGz G.={KQζ=mW[M/4?ԆZ f2#1T$L;ruO6#,OK6n'_͉8X9n2G1zC`ppW 8l4v4|x/wȼrgN*2o6\12H {8GL表 Igyoϖ56cӳ~8o$8^|lG(fJ@^>Hnkcn|FV|!qpriӱӾҵDn= I}->fŢQJ)u {-$^f: *p  tlK eȷ˨ѳ'ZkW+q,"/E[Tp RI O,HG#br *) xZ9x=k%e`@]Ig8>}48bL9Ԥl \a,>$g#0[);G?{NK 4ѝ^۵CCż)@q,wrD:›s mO׎GV 6`ot_ ȖEz?E5-컥 e]MpG$ oXI2c'La$9 -j4蠄n6[FzY g3kLfYs>?"?+oyƈY|G7x7k2bB8#OT=hU,ap}_yUbQ8d2Kɉ~nuHIˊ Q} jV=E\ (޳Ie!r0p|ܛmqY6Đ[):vQ$e8/#QS&X܂ l4e+7q'6Jbn09ҚJTAqxdU9xcLGM# rĦ$ 3>o>W{#khg}tҐrV(X9W ӃY MA,eP-cW_ҙ_vHk̢Gd⑞V V/{5aQK%=^3cGؼrj#rSt) NCMxhs?tU;64xizO_ZbLX֬_L<'PJH:wE1~X 0_jHR3npD(ZWl<3\P$8;5DXlr]i)opeykHUF/<;v\@ܾu) @ !pBH }Zuh+~_$8/=ϯj8Թ-?# Hc/85<|^˿9̣p2g5*ͦo/PK:DÈWB3!q-abl Y>nVxkUcAb$7??Iҽ7"u9"u!F_ @TC[E%{CaKwc6SRp~Q#wG0v1n!.Ֆ:-{uݐT`K[jfl9j]Tk+QkN]oGW6실BӳbBoS@)υ*vg0]lP5{j:{yqj G?& ǽ%y@}`J9>,AK| !X\ɏo~Hf>_Iz$nYl[3[5xe2{) pq5{KG{5KWGe#v%rޘC3qzD|O޾@ޔ) y9=}5]xzƒ,#r!DUf$_NDt/3MM+;] C{`?9^U`,2"jw1[rmdɫ:4%V<(qo8fgq_< 2s&ͪ|ɹp^qBqb g+k]? ?G›B"WD(RtqSfq2syjTHC1%m p"?_JoTtհځ%ʵ~~k eṂT~"0:1&ϥC-^{]zR^zm]ԫ3 ]xq0H+SAq1*zſ> I"y38ѽ6g!myx:tlhB,b(4T+{onxܖD]k)8, ƅóxTE']0\`nşQoӊ\i!9Wdqc}Oj$;ͅA ŇI9wx p < ]T|=WyӾryup*Gs,igT(~/x%8gAO&b/'"/蛭`^p3F|Bi`&vM->h원b7exh l0WZ]o=~‹c]rL`3\OfpCz@'%Úg$@BHk֙|0jmi\)mOUֲv0ᶜFq1{DѺ 1rS:u+JSۤd]T(My7:Fa_13yF*G`K_:mSQFPYWˈj |ౣ9* UW3Wor_zTo2Kӿ5^O`vȌΘ$.]1 TJ+X@K'TA>A QfŻPzT铔'ޛٺ&-k1ZNY [ I`-$<Ș́IK8g]\Y|uP;VBN-lH/vH]xLr>c㨶IsHN IƌuTh nLrR] c@\";/ݖ%I zh N05IlaTtJ.`ykXN 5u{Ihg{[+O|4E`u؛?q/_]{+|hc=݇^ڛ+eX{Vz7^FAZxfV#B$Q.R̃@q8 \Zo*Pu$_UשU<s ]vNR@7"4ytBji`˜V*t૎^YU֝[st&]❖oE-Ƙ="TB|]by}L:D* *0wHsIM3KA\nWw%Q= L՜{c{#' {)`QR1êRGGKkKw-խ[Y]9Of|0(u4XZC,\wN0 H9n ֗45BnZݎx.o^O~ 'UO<$s\6h/Z*Qvv@}^[ؑx郵_J 8D;e=BЎz9Om}V\fr&S%j?$tea:^v$9_.u]lMfs-oHPzFm?y84Gc)i1]@?ޝ7ϗV"+'҅?l pi*1jIA`%HFZL#\@Y9%fGRRs}g/_ٴt)-C,iUg}uƅ7f!Qk4c{5FkX&N2&ZQ0. #bGAۄqX[3VH3V@34ˈվ5cY g ! 댍XM)wsE,JxEm-g@Zb\ f4f4Qk&m;7w}*{c`={Euɯc \e"Z½eA =?L e+fB~aEzT#Ʒ*%kFplG\)93F]p 40|jp#&#CښѴFѴ[?SxO7|-辶o&֛7]^ ~ޮ~0IN"Kp3 ρb9g: AVZZ?.:[ "9 @6 W+0wB} H)^vK`&`V#(r[7z|qJM0zm1J \39v* қWZ:F q,pNÿ>f÷[W%OE$k35Mg@tˀ,ޭdSBH鶁bzCWM4h7#?h5V TleV%RMraJ/W3P!PA"GG`:/xv "bEv>7פXja!JT_kxAy^N9B_5$Fn 1W=3C9e DŽ>8^@U#@9_h]z\CjN4}Ui5&|,΀] 2k3i6L 9^ dV'ۑA;q>"j Z5nTYݶvO)#o􌓋`RoOB0xwah3QB"UagEio}Y7_ŭe|b;DL*}L`F<;wvf3i% _%QO|vfa<8;,k7:G_T4(q>cr,VI ~=0Ťf38_ -.fUSTܼ~S̓y(o:1*KR{7i7FLn')ڮSJ=hrk?\gu;0 ui[˵Wx!QRx 5JMꈑ2mk,Qj1U2ϨIM.clӭ'Ĝ,~B!I\ŻBAk]_4!LNHD(@$Dm|ϜSAr"03Y+eX+޶]`l;j'cs@LpBVKJqC @'J5T3&J (%! "zረ5@r6p])~ ĔO2!(Qs)8c z14Y8!r5w,٘&4v*eJKOA‰-J0ǚ+Dry(2oR#iz;4E}m6Ƴi6uRB5nRrGL^K8MqXbsT!2VP $m"r!2҅ODgZW)qslKh&4d ݞkeXElf~T(){G%P%;Ƶފ3Y`}ٙĩ$c@X?Pgjg8,Ơ~j {x^qEGAf@R;4ǬSHs:@Pn魗QDQAG5wD:!-=7JgKL1 úP'gNZic|(gTh|HEF9FQTC(6|_y|qiFc$JDz`iCimCd.X  h2-18aÍ`LrtG>q0%Dr@ѤEdQrsA`޳6n$Wv ~?備ٻAfr Li, I9qjRe[(J/쮪GwuHv8!]ɧ;Rp@{G*ee ABΟi@ih\XXQI*c \(6 BIQaIR)",S%x$ZH*E}^R10}z (#nTp\`i`ӟך0Қ1ц=77j`XBaj/rb>*/UL˫cԧk0եO6.s11cfsnvoI/™m_)5 #F)x1Ɔ~7&IC,%R4 Iب5ZoIFgޅNV*Z jAfŋb_vZUK ƪ^PS2ځkNZ*7PIHJ'nh7Xz Pn` RkWq{ $9on藦`dg xfGPhN>9P68$VsE(7[7ZmwHj͙O0&{T+C$ib$a< M"jtk,bmwċ [JS9jfC87%eE |il//f+n%h:H/֏镤ԤGrED4%j0H Eh悀0M@p5DTTR$J+#$I0AĠ7'ml(L&L!B#Y3ؤDaXM,%1x+ 6!&%R-"E XGR py#4)$ b0U si12dJ4 g:$ O}ZF7`q#`{%헦`ځNhw.oT:rAxĎs2J"8[wvڈ>C+:U:V` &&`~rY `N-ہ+X_d*UAh'nP0W`7p;H4_(IAOZ13cxO 8Qk;MxW#n^vUғoZɎ[{dǭ W?VɁVH?K8-ՠLR 5ƂO QqOcՏGfMfeUOŁ/DLl:Y<8dlP؊#\g |}%eC̐25tL zT _3Rɜ%I~$EU:X$/^9K?K0.̂ ގҟ C>u:D]9{urY~S cTװ~)LS=m%"0=Rh"HoמVEE aۢH `2BN-v2֗'lF]2cfT-ꭖ%kX+ic<*b0=sqW3PDHڶSSĴ0ubi\nN+m=h8*kMn'2bS0#ԷolKLM߹K2ʅYp>ٸgD3qY| #HOM9:AmR0xV)w["€燷]hV,K%aԎxcq%&;Q1q>]\Nf1 Jx8[^7_ +o;O బvwJ/QQf9{xmqo B/FW RE%4Pꂀ{31ǿ~Y\\׉WVNyh11F!1`Hj>!ލtBBnD@o!6Mqed:UUI9 ʜwx7`yWT&s~AIy5 E,J&^m0!+ėTF-؅Tܖ%ͩzՅ (B߾{C6j+@ qĥI_˸ s->F\ϭϧOfxOEgBK-/ _8O~LDU\}V3?v :[)mGS`'ttbOi=Љ`i:J7>?="aV^u\icZ~wx/.+ȓz@9wizHin4hDo22DԐ;6G/؊xLQ˘Zzdنe~no@} SEshSFעn;y^JU+%%'eݽ G;6md(j3mĴjcnQϘ3֑0AHדY<c|9qk[ O4Cx: y,$(@܌Y6`w nl7P ~<$(Yv`g2-*p˜Zm&tUPsEmv\$Ԏ\7jX 0jؙr`|8q!X~YHaw%cH04KOήw'|GO0ܼ7(WYv= _5o&+&$N~vp5CW677cw$?`i!cne5Ӯyt48N(?F]qd"́Oa֧Y/ V[Zhqnͬ:ЭW!mes, ~p ptz0RderܡW9FL΋iOyw0 ^*Y@m;_Rv3qnmI40ou׷|2%a'_ϵ=ʙ)mԆ:ǀIZ%r0CFQߺaw#_8QX1C&&Jx[ws[mh`'-vj}ZO("NWB@4J(6Fs*~W_-2V_⮖=ugɯQpqƗ.V1frBVAwMh$?jғT{nVZƷ؞s;p5%'mCGwZ^ Mٙ7Q&3CL@BRPlgrw nvN:{O\@+*9=an3I4K( N9r +))@<SPyqWKfa3oGPbO$-ڀ)>yv O,-ځ{m5x &ypWX^x G~ag,GYQӐB[I>" O_L[H'yT拨"/*,m+QXxyt`vkXv 5\jq"Y6-FzNJ2 Nb/$^j8S-QXfI4 -؇GٮPwK-̧ݸzh,gMS,Q솛OPFsﱆcdP46VrϥAm쯣]LSM>$yEחK~TI6`VtWmꕯ߼ǥNJe?g5XxQat/M("jR^)臇FV6M_חu_% H#!_GGr++@* PB !"|ˤ-% ]^p=Կ)|'OlYLgjC9W }('Snj],_1{b.8? Y͒!M !#CtA4N̶5;^ok-iq8Nsg~l8.ހ+' 2ڴ$1HԹ_M{?eFw#bHiNJdoJ`E>+{:|uQﲢA[ 7jGltƣwT^ N-ܫw0ʃ !* 9/|ĩψAGL݉cmP'/$Sz;z&&WKZ( %ľR1 iD ; /,˺pl v=mžB#'TCZfHP' J;u[0<:sb0jX^aݐؖCHH$[oya^JWG7 R= K_oI|Ø/3iMT 膃1á=7I.+'bt 3DG%ޛjh- [- c&ye7TfuígOm{%\rσ>J{'I~]_lyӿIv6O6'8UOUy7OVbg~>}QoVmM\^GxJSQAQ~̃Y͑U7KcW' ZΫT[իoV{QF^̜^(#aIe?BA8ɻ/'1ZqNJI4hrDEp"q-',bH| Y\LqB GDDKlBh@S&Z0_Z+bbG"J &\ @>Y/xP7XZ#ճ!{Bv.c .b:}tr?^% 3ycƥ]Jf^ M|G^5Bo[0٩,y֜~t^? [u45b_nz$>(Ϊh^{(^=PT@hB3 f 4+!* -z ] )+")C X E# ET X`谞3Kf309,~GL ✆Vjmc^~Bb7qXGs]{K: ڡ1vm 4C ].]瓾gJ=>XXfƹ5$2g1锈 &R0cb) : rn}&QT*0KtF8B!N׽AՕ+>|x@-l?g֎$waZ,YLl"Fas0;%PFBѵ03Kww]ĂXhσK*ᯫvUZnkPo*GȢnY?k헛2k.j̰vGL qG $ ح].XyJ Oo_׿&03:wԝ}3 %tndD.21)%M p6AC06V z_mpUEEmKΐQ +%ay2viCX4 #*mVni?<޿- 1vBe{H!IX e 6Lr<א+|\.&׋=9rPV19 ĕe䰶*iL^ͥ "*re1.棤uR;$4.#CLTLyi `ȣ ȇXVzI1?sBcwĽ3NB>6Y "-mD$Rt͡y:ȇe{ ;ۂH}w1YU^\TyIT=T9"4_- r`I1GtkcԪACEʡwfc-N)Z9n :hC05t䗋lgCQ1/anT ʑ,Z'Xt¶V'D>PmoLc!joVDgQ` J#Vӥ<*|Pe9/q>@Ճl32"xJ\H8j Pk9%e,xY:򈇬 JMM2<<@Ki9P}pE)z*J<vK=ġ1d=JYBt-njDŸ7ǿ]ݬp*8K忞6q7ea3i}/Aד"_w_{#nw[pG}؂[B ݂-( sbzn?ŐwD5o{Wc>4B"Ɠ?d_}Edܵ-`wa[|.],H)7aqZYguT1 N CxPGWp<⮃q?* yNJ%F*e:8cþh }Q2=Te kKd7K"E,X]EsM$ aѐ6|.["=< ):!@NIɟN Qt f 1ȇѿh3\&,e@B 񮓐6@CE,pA46O /G[9Dt63SO\?=TDdQ<Ryy Oε*1"hF>T䦜ȷkzWRdl8˛ghǹGs+dr?|C03͏j>e 3Z2CMKD>Tܒ=-R "$ ˋ)HEYX x1Ng1N "X痯U\GuR#JtDfIAQPG D>T`{3>}uTaKB4D\iMS> 8`M4"p51cD>p.$/rWh1q$c 72JJjnBF ,ɲ ` ܃s$=? q3WQMDwnZCG\$4OA, KwZuqŪRfH=)T* D`UakE" G̪"8gGW8D°++Wi)/D>PԆ|߾P#o|:Ƙp wz lCiIi=xU!`y޾#nX^ZP%3/lx==L(f:+?D[LXILޏQ)rPRWE2;Ƙ-\ H]a3eK=AAf锍>I}!&l"-kT29)kRڛ^ b*x]~DAucLI=ow="F!X;Ƙp-VZ2'rbx*W/OW”*Ƿr6N-pM c)_A>X#)ϕLD>rT+9BC-WotCҝb0?ϨF:4RM뭑&\V'-PW.&ذ9^D o}XcL8*r}=<-h3 T{A&\L$Q%Id~n{T#w#6͋dW%*e~y"Zsŀ+ ̌(ZWeZJy7Vdbi21̨0P'4`cOhqWxTP-+W bGcLxdd}E۹mv_n؏Dʞal'2Z=y k}h[v"4z`x< įT l{Sc6^K_XJ}[rʄ=Vzro .纣&dz6ArrRk3- d0ؤʣJGbzTQudZ:hRRژ޵t>s .Gs B /5:"Sv][Nua F;naqg궷"k"{7{ҟmHc]xhv1bĦ >Awԇ M tqRZY"c\g2F= Y<)-xH d FRbGo֤TS, `Y*O|g-2'ByV45QڀQ r5%Cۯ=Abc"gt:5%A}E.JOc\M%tSty"*UYN M}@#!nd "*[=bEx vcAkuPc#+١[ I 4%9ݳ7QȐ*[4m4"xO_Pz^n%g%Y/1xBO*9\S^)9*/|__+'Xo'{\!f%JC kkVbH6kk߻KQ-TH{4֔Y+*DBsUW&+aq֩vA׊\l".{d@;x<")M!ʃ=e:㤭|^B -ĠP|//3T`JE?,opA9rxu2:OB> !VԜl?'^vu*.LyA;8jf`h* S=(T,~( 叧s@>g! U=݋7#ᘆrTarܧT#=25ځ JkUMt*SѾ:@׊櫬vSAdHyS<8U hU!^G7Y` `ȇm k:`퓝?lb+.͂hcwu2^,(|?.^Ӯ,㫙=n`4 UJ9-<5ڀ kE·s,Ǔr|Ȭ'>rk݂@b;iL#Qh")4Ŵ 6 Z!deAzUɾ!ĖZW}zew]pAf7/}ʢ| Djm5!xFrjF=`_cW!V0AO_0*  c8+pA5\ L/xLkJx,_$į>~_+O˫hFw&W [UAcB+!nD~)P>U x֒#Wˀ˚ K9da׊\aiCǓ}KT5_wV:DO~v{{uJ1g)IcP8O,z{zzx^,bbg(gv;9텿"(ͬlrp02Vȗ ENF_P }>qe4h9M`.?twXΕIJ:[yXFvΘ>OX.熼w h0A%p? $ApUٜ `rp~Ѡ_HF,?6⺜|9W~2_`RQcQE\K'8TΘ#Ѭy ,sJ\qTEg3Jʖ\:Dp  }fP)H8KLܮ}ЯڠT[zW|᷸4@k/ % k籷Y fbG;*-T-l~~~ j:x ,v/v-=cM1%XS?h-T 1p4T0o l@=T.m_~留s$q^ۧx;tu}Iuz2D/LxNKHS x/!k,#WStk Bxl5"V?ڴJm?Ss{DkAVߪ Dܥ=0-3) 4bAV* D\T 5hp;wiM>Kut M jiW?|YSӇt ;.nc|ӕ:6 񺮐 & ✰SQ w.- +n" c i;V*`'D1Mj$ԬuFWT!~_ex@~5PoAmqr5x"l,N-(֚Ks;ik?Z[2 |vϵⲋbH2E[ML#NXV Q}t _i7u4Ư5.UF^iɇNsS]~ ލ1g(5ѷk@65Kwo3ы\J&AJ'[R}I- +i)^b^zsnS1ԋiq/\X8xԿ~e x ˌ.*V8 $ +_C-x8;:粡ċ~xB =6nq.5iE>t_Sdux&FV`q (!l貗Xn&שb_zS6}CPdSJ6)y4l&Xqw6|o$E82Ql 1)vH??.`Po2~ jm4$g! Lz V-~ qw`UHˠ4ctzؔ]J"dJnvHi6K|<M[\Nm({pI3iQ4Xmܝ=d5L~ 1d(.@Fx奱9I⩐5캐aFq!J峔 41զB.8SHnk4b"CM !ĝ/xƒQi;_k/]Qg?qgd.p#&"Wi!:<6ڴguG-wؾ|mL[PS Bf~Zǫ|Eo?li4Y0/ߥyϗ҈4P:/^ƋBR'IH:ו^cNX4>V{7 v<=I~1iLOo<QL%?(Ʃ_S(E`\H|S\i ~n6J>n@ؙ2~t|+,7Z>r?j0|?"霜7Jh<-hs5aGifžהY+*̰20\E?M?=G 2ZQo^~d GBXB?}sO`6ԛ .!jM4YFOP '$DHDRhp/bE6nQs7BʬB.f_ OH0ԛxN}vo|w@.v#I$HG|@<М\ ؋])?E-BF,,fFo]Kevp '̇PD?!2P<}R 1Qkx?SC=$j mRBs=.~χwxByΪw Djf19a(gfQy\=]Ŵ>l.-2g-Ӛs絁EKJ*1|RKo#jvyjh|h2gL#|>Z0U=[& oǓӜo`RW. xTס Ƚ] <2 ? \tOHum~Z6|H?9.3 hpYƎ٢DW_vM‹Ǎ~oa/Sm/V v\H:.6 ~G7▩qꍋbt~j-?;Z\ݰnZZO+zӍpWڮj ߔiwz\<%;T8ĩ\v<Z~|Z+U@AlqyqAz\KX˗/bA ^1>6eGj:'1{4XbTZeF"GDh\]A%KKNgj\IO6T}W+%O9P|VX|Rʼnǭp5_lj}X+2$qLoŌ@J:;t},JH8ƫNҹB+Ǎ>٢*3&3E8բ}XTiRR7ӕ-VKgOVr fY(@{NFRe~`e|p}L\Iz&fR>{Lڲ̜Y=0sշCKjOF2Ej琔laxzol2#r%'Ji-G$q a㱐#ףI'1u|`EҝPkCҝ"sҤI[\dЭF$4RoL&$ŁZkI41%g9D)6gX)g ID$?_6p?7Nbw?$҂87i9CFx̧Q9X`2 FATB/,73ͦ1l8S~_u*lP%?GJz':b3 j=\Bѱ\b\+04M}"Mu.'e$H:/BU]Z:iz}BR\B0y// "1)H@f4&]Sߑ/?:<4@ah>R(x{P\0&q+-e$0d#S ZQzu`P_s`"x98[z±Q빎J2 au %0mE+Vh 崔 O+=Ex_HJ?U189,6}N9A? _D(|p !z7n;-v8T 7o7vowP qH,CQ$T_:k?ޮ vRO,s'/TzuV2?wc*1?S:4 Bxy@ކMGjL"SwviF#I3p3r7&KKY/`uʖDCT|db1 3v6#\ *Fs$%9 ztvDZ[IgL؋ ;|F_Ӧ?[n^덻\WǑ7bsŶ^W.tďȽq ԍ^c±BN=j1aR@9j agbGv\, ~PKnknSSZ>SA5>z:-c1HC[+aya$bt9(%v->Ru]Qf=:n*{#P}O"YKm,qč`^B*btq%CJ(CJ޹lUރ>- 04d).hJ.$.Kn}Xƶ);:kl!$ԣQ![f=v fMq'bA;֟4gAr%BIsn }1]Dy ?+_qO# eFwq?(̼Ǵ 5ׁ'cġq>Zχfk{&}0}jT(㵣F(DL *&#`v6V u4);c6œ$TT^FDxcSȹx9'qO0$aMiK||Q"'6EcM3'7VgR&jq,Q!´XYXQD,o! 埃RncDct=IfWʌ1&jKi}?-t3ӫhAZG=3 /e4|@Wѭ[w8ah[Uh+y lZTzu(ޣ06$T{KXGN| >t;T#cYD(A[$\7E݌XB;^R q,[N)]:($>"cloCi]gG_{. cJcQDa YR.)_.iW 5){m-rSAl?>r ii9GOw@&a&Nx{깬zhzR}>l0)gQCږB\s@2aIz R(3RDꖲ3"xVa_IKdpzC>V4 5XW#e0u2%axmhmos!$2M r?ލ:-ݗVQ(ΘCm~t]IgVVqO *)=S3^e4Y&bēľ{+n'n]q3Vٙ^c^UJd\)k6:٢tn5zy[dRiP,0!cBJeeI&[^Y4Ƣ)©â:k5r`[Z<&&#?Wb9*jQ4N؏znx1|t(JS%Nͷ^xW |qXĻ_B+Yu^Jt56B -Nj9D|6./|}r ; #x5o s"ΰGF )rRoݕRIJmcw% 7CvQ7w R_e [D3/(NpN`: .7La>s\Kab z@2ЅG(bw>E~0cJ}k3& 2c}XϟaX5a,`?[p%EP4VO;SO) ZM>48boMi/>.== zG_ʵ)2`kvItnp&˝j.nu/wp}=|q@rj]Zv㨶O-ܜs;w'Ut^Kg?V\b[0+Ofzpc@LBMg wĝR퉟@vt}wX_K00]d|$vV,KT>c鰬%tKew9.^/>g&omZ?kWެ6EBgoCpcVnck}oG~>߰_6@,eb,ms߬C M%Jċu %A41 g uATʹwjՀEs27?)mh)H 89H"oƬ<KI^)-KVNcX#-ED )@$8L*+fr|HU4jݓd=3x}SВGgC0p\B}۩ CZE갈nEKj\*q[ؒ]YykAN)V-.!V*V iƞuF]phSI݄52]9"հQ}E8ZdkvF,vDk!+k!qfq>^19+Q7m/>uNmU'xt Vk i 0@DjgL?>=T7#aY$q EGOM$֦NUoxf0Z>wK{WʬE()yߠLCzw< ̌icGkSqk帱r0oE28:+`$h,@O!Ę !+"Y b7uyP7ru_ɘ7zC8:!)/Bi(MG_ X=v3pe$ īևU5VVF]}ho2u w=<~e\$.Ķ1Ez<\Yɲ,_ՈOҭ*;hzg#4*St֖;+>9rk`𔏦1"f%(_f]^ijiA;UDVLD \J]W"UǺ40SU{?xeD r )d g[ Ut$&"ݤ1؀tFep  zCc*LM9B.8aX,P^sgdikPY>tm>89[GhTG odj ot4ͯ=j#`(DCĈ`j,qӈШ Er-NȠ=>D22/'Mc__2oA,~zHSH di.iTnR0f4{n6Y1GhT@,~HFFGw롹T,=B28|͓&ֱD^*&0 [ABhGVFj- A.(GNpl8Z'd^G['<6O ~M2Z\\Vսa+iJ3,t1lg @.P[oc{?`ssHex (2vo'N&xV8'Q]>5eQ^\% \hr@Pvnn!1X f.lT"y# Sݲ{ +Nf р+RRbb61Tn%t!rhH7eo6ɐ yN(b 8!( Qe=4ee=^SQ?6=*]Wt?Mj'4K^v9fH*3u~Fep ;gބ篓{CWF0ageTQ5OVr:xyk q،ZV<%!Y=Ch-sR3X>8+A{.)ս3ve0DFۤ} C*"˧@M얬GhT9Ϻ]}lnL/W \j6<\}IU 'Uڟz21iy_ۯOm8GhTf68p~T,;DSsRA_aKh2\K,,1r0bgYaÞM2T},TШ eZfG3% ǘ7X`+z< X:IV!AI(Gs0i(xIe91E{8 $4HD1 EITF]45ݢ{L&1QJ ZX:ka ]ot _mwy0|?)a=#GOFO7q:j>`'2ލF@)峯~nwz;=M_}ZrZ;lr!\5iݬ-;amڹy' ;f9c#|sK߉K +m~Ws AgǮ]4yL'K`0_߶N/p՞__/bEH_TISbf S0),:;m4F+dnGݣ[ûciFYq5+<( tӶ U0G=-Ci d[.X0oM2gh<2ROYmZ dALad*n 7XDTwv)vYGnI7oס./W)1/AaBazfl ©Fp"bO1ZkPOT5M9S]h_st:4Z9\f=C߀><\-`ݔQN+}V3[N-;,z#jz+s4ExhrZ$ ~F7;<|Y| Qa҇Ulp b"g 1A)FQ),MOR hc9M#;W#[/~ibqxXyrynD*7 U-GQJmk-[{%Cp+j&]DÛbOo7V1J?뿙bM1V} 0lfϛ &d%.gMh6w^@B;#! D9Ϗ_Qmnbz+rG?_"rbֿp/)'ZD0j?ϐPX)LjPNpkxt^I9LFv#`{&mT=p j]aA=[2\BWqa3Ê=ىg6B藋PC@~H\{@>eoeI\ܢ.IE)&ٙά:u60my _&|\oM9Fr 塹 /EI׋,X]30Ig5o}vaeZIlJx`F0}ڪlWQCe܏ph0B t*6$,>cc.Q(ŻVk=]hti )u$@YhmdXA $kс V*Iz,mX"poz3Voo-+fdy5DiLx;?zf^OW(A[;Ԯ :M&S?ٽ5Uio~D+ED*&\BiZ8pOgfͬ#Gf嫥.j]-O5⎗d&atsUly3օ7f ՚4 TcHBl TÈY7Pm|w#.(ͶJ,֤ŞX`));bѤ`M,0% #oS8Ţ6N!!M{FZfpM5̆6z+rqqhEZr> 鰐`F8?@| fEjjt478Dl>㖅)櫯TT\lTcܦp3N~LAT8@ϼW޽ zeah m]ZEa5аgyjuÓ7Xܔ*Db0* +q K,gJž!Mۆ;)'zuR0|vtD깅0@mWy;ZW RyZ5CEb"V@>eMQ LSt;ZLZV+{&a?\*&ݛW= Ne"JP,0+$")@βR>d}98PFtų2I/," M9]1xe0u8O[Y\ Xn>bfҊ۷QYfObpeF󠸣bR^(z '[^c_zݹ/>E|VTAvU5uDI*3dC)X"uJfc2)4N"J)*#)LB㔛4Xݡ}1)W FXI]b3aGMbq!<p^Ej|]Mֈh9$+ӣ>+(s٪ISb(?{Y"N0aI(Yd,T*d<% PBD% AܳE1;`گop,YE82J)@8XWq"JT&xl9w[.W.woG/ݸ?[,`~=ʏwjG2L.|(o^>:k7]3je=+>X6Y%s xRI-M"i+TJFQ$ @'>H4#sI /ddf6$,3*҄Haf HRXLK4,A"oʭH&^ |<:'̸8!Bd4z\*ZXl&:B"{!K .4H0&9Fڤ$R֕跦ښR[n.}NeH%sg"M%ȔG1JamYD2I28Bd,Ze-G|[YI}f=֓oW(1';~,dZ#{sꚞOWBmJ ): ?|.q + >qDbpG̈ =f}" SDS,&h>7W?_ꛑgJ 1,1H2qN,&*Mon8dƋ"!?LԤE[WsL#Dn^?UiW! t1oAeI` _Nmf~}wQSks*׃/~t'߻/Ϊl` +:_E_l4|Ӗm+N`<^hg~KCnkOo骮Rۍ䤶*|Mx$d<) =_9.vLN*Q+A{-'Yk\ hIs̔2b;>܉Vm>&y_h5i?76珃d8vgok_>~ Lٷg ;lѳRUCL;5:>t= vlwL׫D~&(X]̼QE`CP0E"t7IvTLKu`(.D2N0V,qʨF R]δ1}1]aMX( /ǝVg㛢m0|gD,(A1:V6TLkVLH HKsLS(7zڒ};;5;5AvtΚsڄՠCT"7jjEP 5#Jhڒ6@Rbjk0avS9r_9ntG03vy݀vff:^6b?o cX0s9JJ֞)=gFC}YgQVQU裡qQ0NS6G8 KbSTJӄ%1XžZkNt]c!*CΖ UbBVVFjI!Ճ@*GT?ʤP@m2EP])@(9ʣȤ(< \tưX[OJpR*HKH!|ڶjў8Id}bۛO90e7X Β߳ Ֆ۽}۷|IGzwI5I7!| ֞٠ S|X$M4Jqjs2E4֩p8%G$Pe93iBI1%|% vB;l,fjer+ӱooW yW"č8Vn+{伪/ o ѯ`}o=G'$k6U޳02ku?/:^|$:O+bGZ(C?|I)hA ڠJAqJVnf!Z`g͝['й#^tD9alyd&{ 4ػ9zt@iď* ޥfykԤڹGI'u;)=OTd\+PB($0ilhS,Üe4UTk%fk:u㞁Ik9Z$;f0pcq~d";z*/~_/~*v~_}/_˗ze|e_Wf|e_Wf|eG\1u, 'Y`dW5jayXW'n`Eb X04AH)''Y4a2~ #1|) _ ×0|) _ ×0|)Go%Pȗ/9R/!UHėʗ0|) _ ×0|)R]KR! `D8JJ9  f\G 0I~&B1-gO.Weq]‹"8&֋ S3D0RQ+I@VM'8ayn/S̳?޾9}ڗƍ/?p_<:]_-J|_<:*h_$_|_!e*~ǥp3Nln?Tg}SAMUz|ij ' rqc3]]M.'0̀mW 7N鎠Qsub9K .fi ЦuDXh P,#Bx۵ -ṁn_3˒}k mdfP44?G&䡆}0XIe""{3hIpxJH |.`I2uXQGNwG#фU,?{Lh7`j12pTޒ*Q3D ?gիoS,LuNo bvI?p?hvRU1n1YFXkOApkKjZRmnճd4;*mZ0̡h$um`'m>Ͷg괉UTfwA KȮ*CoDH lt:[K|x$h'cct.WQtЮkj4?ح>iC)4(I=ITD5ɋ]j,Q.Ä.XѮ\Y9Laάd۪g!2v.JPϭ=;b-u÷K~qzkxMkh6,$rwLL{N'Jd]=-λ<:8A݁Ԥ., K`#ҔPEa?ŧ6g2y#pi匫zGӨ.X|fͬnrnKc wWe:Hڽ,M&R7P[ԍґ:Z n}9 Ҥ?Zh"Ki4SGنmxƒ%1ýS˽I=ZնGS{ّdJO4ˌ=`3׬ 3|jIag>'Jﵔm{`pKUWqdFFW|سŻ6 }hvf7xͷ[̷M7mnvM2n1^⛯q6ύe GدכR;U~^|:uҳ9Up˫Ջ_N_y9[]5چ7~#vd8mlc/?k?'*lj* UV*T,6woN-rެfa:ɜT +Zl$+-eף61-_!aiMzި3M d[xe S;lK2.G;**Eƪ1fO_#axv.LoxİdUٻR4[n.._!axT :殍1Y㊎S.5wQO)SЕpQɊU͔@T|} 3;ggA]s/JU*40|YB ^l5E켯ɩkVFE ~hl1)YD~ƭ(Y'0H> ܨ%E:2+r5~ 3cp '/%jm5Uq%^B x2䘄W69YQڪJs0<0ޚFF f"T9!GYB ZL/R2F+{ S0ފLɱx)i3@eaS=9_#ax +՘:0d *R/ q~[G .ucuVeo@Jy S)!tIiӘ{E(,P(kHlo (=($!cEʥYcMZ S0;ԹjU;@Y6ptV/ 3ցTo"PӐ{^B ZᬐdPҎ,vP<&%N{m q~!$x'Ǧ_q-9Ҙz3`lzHTGq+$AW}3v\7E%~vxE[EbޚM F>&nZU! b2k sk$L; HfT1; !~Y O aGւke%g |xL 𝆩09y"`m 3f?|i),ʅhԢRGψo._!a x?}A?jyCӄTHM#O_#a^4j&ًVoDVMua̯0%*U[E)C$ WHޙ$'2Y *u\cRg_xH lnU ac2JKWH>Nˌ Y^hXId14p൛a3QTUj,l\#ax8|pMM22ؤU2natOS_Z[#axgibB%r:W=|U5;j- |HPr)]5;* i uHG7{^"j"3jVzqYBVF Ax29ac=. $g#2[|ףgS77(Q{~Nqzq툗p+w39q;}wzkM:;6FMNݛ?/Q_Rkr~9XٴGAp?0#o'_nz?)2+'W?|!zus^`~/UuPlGWŇ`Ua7eà [&),(~sHMHW䉍1 ue,}bTmo6xBa3GvVFisW;m %ڗ"`T[單GD|ӂW/GW<|]#5˫ۚ|wz?}| `!= t{IڇUZ'TW[rCs_.=qt-8=rqś]iv.ө>|X҃:'= :=\OӖto+&i^|J!/oN[՝lҏ0۞zXa<UQ/Yb\ Fvq~FMÁCx|2s+ ?~}Fʅ|A#ĶC9 ~M[z(mCo삩#(Pn; ]dd>?޼*1ok`vo}i˔.?F.xov?Q]/YmUW=q!]Bz5`d-'FF _mDt׳jQ甅C3Tc&TT M[m}ʹRVTcCVLO7LtJ[ >בclҩoEs4>4.Io- ՁB[%kAgDiFʩM6;;-6\Cn@R֖XL^Ae&I%ƒQVVҒ Y ftc1"{NXj!S"w`lI?wz2W61k"Ε=; ۽^;2 IWX9玮!5T.qNe*kR=^bIfg 5ևy@~h TXhun&DY] J°]ko7+,nZlEVn $ _m zk\6u*j5ÿ?0ydH{Y g.3QIaE_7o4BRj.[oe^mFaeUX@Zf@d!yКt/cecHtqmZ{ k1Ҙ;gr a ji8/m)[I‚k!ntÚI`td̵T0lW#E=Abk1>]E=so0#6QNF#?>ŒNܕb7,J6 !#›+4'K5h+{]rg.#<PPy@@0J:Kk={P@ʍ`;ղk a'.@V'bJN Wfk 5l'L+F,eBE5*dG` Coz5keP?yLڰ b9oC,rVgl,2b#c Yj|t\g-JL1#LjҸ` rnboY#{A&5բ!DeM'a Ts806L Fx̖RbF80J&ڷ`+(B? 0po%Cqjk)0j2!6r)C )cfR [0 >04X]X KWɇ(aBtX`eCw JĈ鷛rTMNjHҰ56"׹DkT]w*9caBmaLX`j?KJ!q3Yd֕vVebcM4*2Z2PBwk߷jmL,ӕRWF9&\Z54ȶ6RG ;H &1^],~4*r8L .-BX~)gه`BW1g ~Lgxlv1=9>1'5@qyΦItqo*{56<!u<5J w|SQ˖z'YqVbA`0Ih T3kL:b}9L 5@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL/ ċbA\k ik@NKdc'.@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 eI@Of`@GKd}@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL/ 8!1 4 @Jb}L R@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 0n[ƙ[ZoRjZzݖyA w9MA\g huH%\!.Y x%H)ͩr%qHm^V 1Wl@gOkޤ$] OKw<'Mv:*&3ZXr&J) #i>=2%˷ *0W?/yS(B1N"eOnG[g98]LlpzgE6 j4XTAov7j[.㫕^aw^gL,EVĠ:Z'#o7~E}rURܩqH[AL!L( _")MZ-[}c晶ym=i;qgU.jhg*njT Rp&,뮀{:dIj۝n;(o9?KcU ;o\h݇SܙԆn̟_^_6r48^cՑjrr|lmXVY%d ."\X:2`aڗM*mc+jˡ_>h+az┹/2?[el) l"( ڙ̠b󏦝:UWB#ս~tZի|Kv(wO>dtu*myg?;[>8.{qP>^Ԯ/ x/!KܹϿ}Z8={ls1~pn5o~ѭiW+%<ɹܻ/C*C9AB:$3Qq,gj.J5lQo/wgX}T%KYoٽd ^.m@`[tmǷViBkPmH0il+'۝Z}خrrY[["OQ:r 6P Wi 4jsYld, s8OGۛ8au=4hwi_? %9$Xb䧓iX\rD+[_.t-r=p|]BKcQ&P1&f"fV#,]9u\d@c.=k;WccPPyl]1-VЋ.ꟽIbuYa3&&Bߕ\ҡI&?rM5e!ݷR |8 }|sUOp<)1>$,ϯg\ےARjq$sγ]|aI_5X;*7\ _OJf:yr;岸ԊxI^n-مe<(^"()Vʕ%GuaHrtt#x~Gu5:p@)JQ"d__-YWf^.GljO9 `ߵ7·d>Jo͟ZGW\JYgQ]hrM]Һ.5v2 .$_ZAҾ\Aw{I36]Kw·Wl0Wjqhv-QUt5k_qz7--jcKJon|[3b{3Nnmf]Ya!<6 #XFxxa1UlG,s[[lwNnku[_^:-9X#˶j JFeCPx=l8. ?>{w?|_~=}ëo o-_0-/V8#}5MK7-K.oiwQznҷ Hʗ;N l$~8_|l:fwG.A~T@l,%(JRڮ5!Ӏp[asݻ+c1ҍBqR[^9P7"w~8)l{[F-ԅ9F3+ȹS@Eѻ%MYQR3D(ّ}ڥc7&ӟW1W#ΗrwwŎ,3N?qʫdp#uXt L9F֜s)M E&T/P?@6 uUS-7uغa2 >5UH&E:n8!O!mT7(D jG)g4<|Ӻ%@H˸Z$*B[SJ׳>6-1Q 0sRLW!;_il% =.FE&,98nSTed5c)'%よrӽ-GsX[Y?ǡR+$#=jϜ!Z mXɧy^HБvt76p. }(zi|z)5%m+DowmI+}1?rNi=M \ wYR$9o쒒(Y(,["ŝpƍ}`K)_q_g|gY)OuHT:utT`|)Bn _n `/PʄPibI wKkQvJ#~x}V;ͽ%WҾz0Y3)SO=p cGAVSA )EB ™D~=n_B̗-?V&3ާ8SB;'BbkIX^":>L8jTB:DEv ,UXχ;W,Հ:\)M< (/v0P̓@1Oďy*=Bgr(FZޮbӣUP+p6ɟêu΀&VW lVfjqAW|ZX\1 2YZϖ-UڶZ;hj]~1Rzf 6"VTڰ6\~bqHOXe<찊>O(xhzC7Ut (dl3m@I{ ĝ&y/ tߝH`Ηxyx:ɯ _g?.j*4ͷI?Zncz$В)Ӟfe&:Z Zwdhgooх8tlNgh(45bwv'=&]>OI]RC $ SǐUz}z_j&E&K% w;L㘡z`QKx| ߷}7Eb?ѽ`gxһ^w"cE WV{ Ǖbe7):^Zu[< pv0'WCh<{4J}hG#˼8# ])G^>rp$鷃Bm/qL\nCWjX-(J/)egHt!CIƯ!#nOҲA8gi]t sI&$_/-wVl@3un#0|WqƇb "\e$㩊d*%z5nR'{Rd~m|Mv}7E6+xRkvw4mE֟3?K3쭊56Ôɵ QfKL%GF*')PJm=Sp@d)8FSГRukH8YpcYlVM[~ݲ٩mb0{[l|9GV?ڗO+e#/.sB7!TQqH3؁ R$1QTđV6]%YbӹɛshfҼl,:"#Rr&,a/,KNR2CN+F`j$`Lrf({0?2Mu65`+pRJ:A"R0d4,J ce9Kp(:nf7RUSYÙIEqQ/?#.N?8JZ$7㢕O!&27pIoL&ĸg!9r_ 0Ox!yhΙe|̈g"u-V2۴5mk@.a Y)NQriN& OMȵ_!msUd>^y)rݼnKEKO")}9tt۟cF6:`MRRajE PUj 򘸼f4NoM|%}ntS|p^g \O6fM[-@ԺoFMҝ0/B#t#V C+W.,,o{Kv=پmLN"U9*AGOjԪ2J:x}j?{p7t7u[8-1PEjzJ/_-3(E n\ޜ[O hE&ۢ<üUZ|muw{Ӿ(Unᨚ9\߹(6?;S2a)xP(O -8Oe&@/sC)J)4H!axSc䬎7kG9ڨrJK'A2>,qJ<~4>{840_1̴ `@B (YӅ_? qF]*f.qJI #ؔ8m#Z C =Tf1$Nh*:MR)ku0!R=tÒeTktz0L.D$JL1gJ4 #g5* n>*C7L~\lB8U;1z+Ls^a|@fHMY"|Z,^jd"B]rX,_k{^0=t}zXƾ߃#> egLc2D}TM\އGPP)iJP%\ X=~z叽+>JA=d֢5S?44f7?c1|er;/j/~`9ec{t b}"1Wix^Q/QpXK1ԩ$ƙNRLhNt*3f,AV%F >eYg%W8 %IhF`ΞYY4R3HbI 8ޘ,r]ԁ~V>xt)|q<=#F:ڤs(N˽#JzR1uft &VPnl¥ ?%;jK*0giߔj$Dm/qO U Ūmwv=X-Xa aaIbh<*%ޓq,+_z_ 1.%Í 6anPzHjRGuWWWu-] :<*ݞƎ)!k s)8jSRrIsΞZCJo[Ҋ h)bYD<X;@t&f R{c;ű!mz薯9LBWh̺ Y-X 4u@BD Rncc'>-kEKjNwK'X 42PJ@;AH) 8R /[m^7#=qm{O8߱4~B{r@`4 ~ppI..D i8I:?gkܧRdll\uT]=r_oޯ K]v֌&֍Τ$%#tǨ{-ɲ0n6Cp }̞UÀR-6`(k@U9sWt˫HEa8]\.¹;c$3,f?-o~}6W2:Hd)ٴE'o˖FIX$|Yq"$ RŅ쬬ivx=F]{CfrV,ko{7n~풣Hmviߘ0| TʙbtLiH4VNL,> G/Fw4='ͳY5kZ%D/fb:HH28mWډg;"_9Udàbްxq vq݇۫~Wϯ.>``t+AP{^h'L-OM>5EU˼%W[>/ ڞ[֜(|v2[7ܟė'| Vq/*] l>(6zY͆ȡe*Tn#D'9(ePy),`CVΉxA<0!HW0"bLKZ.0Ir멎hb1vt,{t\2~wMb1޳{5ˁm[,n+sSKt6~Czb0b=oIm< \4h8NӠem'm6o[O0]J<\3}ɹ`ǒdܯRcgm@FO,qJxNm o\ /c8EAx$ aFt]zL-yl,%=fw d 6M ͔B2:M"Ir*Y|^AjR],-C'p;0f/|2_{ 'U4SP= gG,H ~(dvC#F7d>ƇjlЭTuX$`bؓ!wo_TvayP4JK3kl<dY,'.G~3dSNB)KKcޭ:9aКX޽|&}nT7p=ZU*T*.)],sPt*B (*S%)2)QHYu2LwZ0Y$EkbzlRk45[!-u(Kh>uq8E*}l͑ҥrG\T*o-|wQ՝-Cnz4c*Ξ Fl:Ie]3¢W\?u"fwt,yM }iy.Lyax}?y4lQ-&44 CxPBe5xVxaZuXx?50A~,ZDRjfHiMaE@k~RBZ]mM̶&֚ɮ- Jd. V:V'!tFj~*ZK$vFX7_w+7h.zZGK* `ƛ[xb2 ،)$Ť5ۄ&42(gyeR BmR#䃗\aÕ%ԚјUvdp_K>m$Gl5xWyLlwH} ֘;1GViG &Q?̕7BL8ur?$P NU DN,:I '1beh۾-c:yy=K:ǹ IUp;oB!cF@,Pi9KWL/Bk%i[y= ̫Uj7 antT'+_~e o ߪzGReۊbUZh)綶(H4ⶨ^lmr rynB mmȗK]y±_)R _Bv"Lw[BBR۝zPۊ24 ޶ZA )fɎW)Qf)Vۆt[@7MimE]oR4ܞڏ`Sm׷B6k*QaMmejno5-pooåmYn>\Z4j":[[ڔ"Kh^Jn&e@Xzr7j[U"Ҧ.ZD Q݆o>FtcfhjFy{=NDOBQxT~"J/BQRBwX"n~wmdf"46wQEvZGw0|1$r }y(xsgL?D_N.rY7\X|LfAR*J(uԹZ:ܤ7J8[Ϝ)9IjSVh8k58M>Đj.,^wYioD;EZo!6e@k3 d43O)~4*'>=c >!doWL"/]F(5i(~C+X4F06;o߽{c+r!J1k|A{ dɵarV <:ѡ7Ycĥɗ g0-!UZ %td/ y؃*ZE_̛nB >̘6io h!.':Sy镬5.ɷdUW455Bm yۂmA޶ o[-m yۂmA޶ o[-0ӤsǤDSRpt¸U5 \As Ug3V _*@Dז+>O~=a|TSxcf"v0"6e^,Up)7KTN9GөE*0dtZ"aKDŽ O1jveZX^jz\}q:[:?yn%jo~ M9e]S|bۧչwչyK؇żAs ٚ3_^bh0۟x 5*E+C?Ea/9TCŴ| |~(!TDY16H€a@:[TQ?{Wȑр``AlR&)ݿ~#GE)TFeEddFI:5ѴO`Df$PݫG!=Rr_[33HF?1~N P n:[bF;TB#0X(G(*ё\ժ=/APy)<`CV!EI<`+8t Pg1Ia{a!TCyX<=qڭs#V`u ./$66lo9ZP=j/ȵ36'V;NV0ou|+&=3P.W}3F+{3F~5v 11SC6ď,թgK~^ԊtL*JѢNJ M19< -uAǃ i[I<.yb1F: lp6HL;HysQ :c=bߵa*9E]UMƾ6)6܍!Lꪘ M݇_pwzKDͯOI~(d{|ٝfBxЋY_ r)_vc;8Wr JaȳAŕU/b% EXRCA @gI1[V#KDyQoBHm3{9R#srfW& ak9X ! 9B>x6\y,qXbaLz1snH(hxzZsۻ(r|U /[1:6ˮ Z『8TxsIoiHr7gݻqX2^YW)},clGA@;6݊K'6uyEⓄ]N|[U?7[0B>Ҋ lR{qVV;گݐRTHuތernsi:z8K'BX*N Ỗ*7:3uH57ɥftZW0Y*;,dRy<,NѺemZ.`A[-D.eeh-K aY*&ŝ\Жwܦ J0fV&>+z̽ZqKɬ//6MNpARz{3q^;tghL'&zٽEo0.GxlHw7 侥 Oa}Iw9VKR2㺺}ۺ<U,,2 }o6}Yp ~üFnk r)4 &?bٍ UDa2iyi&.0 \wBpg?P")\-݃Ic!a*R7J,4z\map7Sy6%7iQ4Vրˎ"mCAy".UNH\\9!=1mȄ=sQ`GQMp2uRFsaY Qa8 o(rnSHDcMZ &A26vwTN蕝f[tt:'Z]8Q BkW tJ^s!as9q_gS] =>u]٫{QE_3`ʴ_w>rtKi~8i<RCl+X0 )"᧠@P5/%$IJy*-iƵZ.$ )S(o(,Yydc3E4^JJd 1Y,-,xGI4 fi {ln0*ۮ Oj7* ?_hɧ<<HҳNQ$ðT#g$!(IJŠoV:ԇ8?^GN -UK7'j>t0{YBLb**QB9^ߓ#Y*M*0@--*0tCXpb!g=f.[PBHy:^\)̈́7G+Uè}1+J¨=j! fp.}i!Fp1i 4J` aN1)BWow;2nzn&n1bxqZЪQ⮱,!E.5ϻx;Lu ’rroDXǭf0mq:m ]6 \}wRB􍲲FKoV=> %Z. |L- N9ťoW12#XP1g&fѵ[1vuԮ͈ܖekZ5vk%2og^a FQk %i_B:`^e.Dʱ8>0XҰȅ4*T0u4HP&P ZPBXGmаˠcݐi@ QzHLdu4(&T*-M4mдk +ôvZyY#u3lJa^iGj@Y@SJx->cJvd:B///wW0aFn85TSmڰ:X^6ͪYƝ$d};o3ڽ#@q!y{`$m6^ W^k,%Dg}byBl`N:E%~AiRH8u3baz5DcR23 [짥<-VmɯkyǍu=|(dSZqEK^HT}ƯI;uc3EIk뀽"kPq!+>?V$ZkbntgB'L&z !UJʼxM% p?a\ u1*A, !|GXx7JuaI ?OztRʗK9,f#7_v>!eA7(_&EndDoJRJ fVGq8>kECk~R|gF%ܽN؊![N]&`W_YfВk}ӎkxEDуq_=MF܆cvP6#gkh^xcZvM|AEp̝2>y{o=UFʏp_qYL-a\6Ɓl.lۇllefpE>t9_|LҕR;p(9F1v+Mgׁ!{?6Zgp}HvE "j8x0,"ᠺyCRй0\:rE ēyCAfnZ"a!O, 7I^ C۷"rY]YY38t.D>$CvƸ;6d ܧrW B %!lkO2(l Vb!X:VV958ﳩ :g~/Wr73Bgr<UX5 S띱Vc&ye4zl5ͭ$rkjmtjmWLOV%N:_WA$.Ժ8DaSW7AFxzZ>}2*ym=ͽm==/&_rr![NcGuaͯCVO^AX];N^ǩ7ɦg.sU޾LO;.:us/J$$w|Q5f(9㨰]Gׇjy㺊_+c'2G#En5Vjf[G3s\/JI rH# be}0za-a$EV 1lQV=R 6x0 FϫO跓~;//1O&lKịUh\ 62ΤFR :dr,{L]oԉ a8*f ULZq;e\l:Y"C 7SGbD+D /ڷ;e\WxX`|={,Gt7Xkʬ]ou65 )B! "zረ5\c\rI nd#a9$$V8m0Hmc}J!"'^q<9 izG{*6Iyv(l/8\ٿ^5i$#i0>\nG_q85}!; #*osy`L&%"*0+cV)NSk3giUWp) R&AjZa(0<Ggu%p:F !Ƞ)GS Fy1L0-'Fi1f8pl!rae P!uBSl+>[hB];28. åh2XbZ3"|lH%#IMjMXIz_|^ 9qK7cAXqaUN&P# Q[3Tp3Y+eؠcHRI0{uU&}Á'5؆R8H0h^'(oKT&hG)8n7S D$HW[UYb΄875VCp0->ޙ1? (H&ED52]oFW(}? \[4= }ںȒ+I,I=,mQ@[y|fvvF,Up%G r YH!v LÿOW4[O4`| a)w!TQaJ*8^JqJ"*)o7v_g uԑ 7ZJNcd d`3v8)8F"r|ϥSwl"\q1%xy:MUÀp55g\@~jJ鹳2)tp~( Ɖ|Ň<4s;n+S*z`e_LWjӓG9f$,t{CNljJU(6 ^R6@sݚS}wVݞŒgz|Qj-!Fs=cK7ETv.P6z\0Ca$.#1|˦aH063,A} K+V1Y3Yb6f7^19Tu2ɦQ(!z:j2DFÑK.@,EO1ΆU=Ȋ W ˙]];~޼zM?gWuK: ÕǍ$Im#4݀;C/Z547*~hP>60a|:nԶD//uf_ /IN3 ^q/> 日v|\,RŴQ.}B`7s]cqdX#F;78粈ʇ1fS>4)X=r5m}M@wy4H)'?4Gy8;P1~N P tJˍv+ ,md#jSё~KQ'OsM(üDQ!hgQ5& ]DY)q{P/ۄje ٷn=40O {Ew+($VMzxuvV;ʷZgi1Ϫt%a߂|zS"KO_8*8ҩ-/'oM:R_ptLnwC> ʲT\?ͨbNy6 T9a rQR&( j)gE4_ebi8ucIEA2ohAL 2 o7|oeE.ˏgMmkȱ 4{W}T~Ei_Zs&/jr5JwףAQ)S {:)]mg^.QpcvؔԫS/}k=`@Љ =s>seʥ85i#WZQN)93Z1&]Ka}bnD׃2[I\UڃyBݩ1T)b+t@HmojRz\TknXe\a*9EӌŗE/M Y >_A6 J݇_:β߆E. |w/izջ`y%ڗlWBtJ_Z\ZhmNkη[/Жa9(¶UR?ȹeWqlmIc0VmP&ARL: ښ=cڨ(3{YR#srfW& aU,R#䃗\aÕ%ƴ@19 iop}}9&42d'5 {-<>ϗ5ݪW8vhGHJ>"zW)e1+v0 Unq.TPJ|94 -y~{@d((:MչVpb#V4 e3C ZGKZIch՞2#,RFXpt,YR{lH.=hZ3lIyK85緮K^yz aa֔Jy's:ӾK)?^nu{uE?j~^% W^ƃ+TTRkC-T[8Z!C aMb`"9ȊM6Qqn]J1Gl s5~qc#fR;F6Z4no8)e>ySQ.,hT9Ø\\iGPxph/dj@3nD12 03MDPRU=fFqX*6\8%uuQ"70R:XCnR 2z ch00NEF 6@9;}0>&zIs+`Jn~υ韹0N`B: 1yyz^g|nHT%Z08QtL"Z5YQJ@7hEQz'%#:#}y4+d0tӬCxe@0wr-ݴxswζmj+73Jkg'WW4Ah&N.䘙~a{WVZl*z۪R EU%7ih 5(O+LwMkRS壘KE#S:/42E G#<BQ1F^* Rn2GyёaYqGuR.k%O0 ^jkK]B2[SZ9μ -̶}FF?=QiǑHRǙ4>K R@l3qI}8C&D4D j()TMVР|t|M40p=`|KdO~}ӶClsD[JL.p:JIЫ wK7| ,($G5^)EA͸6 SYOR Pd =K|}:X{))qJ7Dd%0l\5rCdЉ, cB9@3o\*YgϿ{S܂|m!xz !( 6 OY~.?=pMT85]X@`B<UX-zg՘IDk1hFs+@6WgqPSIޟv_^{B+.uXF|0arroDʵǍf0%2s] ov\o;w~5OJ~XLacCI"nUlMքh-JfR>jmGm7 ,3A [b:麧tM)Js_ db,1VqSůnM?NTIF&h]h.;<{3|%gS4awϣtKaKVn&5a*Fg]׵;'u]k0RkR/b9t1=͑4gFJsE7Xg]r$x\_}IgJq?El'ɂQG"`"RSFDD b FрG!eLD˅u {vuAխI>NOL5wQI]f'vIf> ̜.䦓UzHY흀, |![1MoS8]d@`q ]hSjOS窓˧wD2;swYw W';N(X(fN&m^mdJ 6li@dKֶ d#B^ 0>ixm]iE;׽JRe}~ +Z<wO js2wmb3o߳NjTrY߸eZk{~bJ)*6QAdf@s\X>.T3։6IW5NVY.lPT8lOx[ ybp+[ҶkTLw3ekh=s|!Bj﹛աtve68$M\j]I+/4xS^w{#*STrut0&1T  !Qu WHlBĖ2 oQ4`QU_RkmH..#!8d9$ 5~J\SLRW=3HÇ4|IciNTUUF5l+ñGED=q1 !Q% 8D LH؅!Th=0`#2\*밊a,%V"w#dG(aԤt4ÒvU(F>|ò1m,hBss9 pjD !j:sNɉ=+R6v@66.%sn[AT r8;ɿ:KAm:p?@E(Vz;@sIqp߆h "ɹF:gܪs&1_sm17z^])%R}G5o" )!k sQd lk .r䉿)ҘET`ZmbD[rKkwUKĸrD  ,PRS$Nq,rИCu;"!蘄H9̰RWN+ĀCd @Ժ624}Zٛ* -X  hN$A[B#jp##<ѷoӠ;Ay  @ $!P`Yh65a %S!5kHu;y3 E 0wCַ?3.QVt}j!|h.K0E '5sjaxc\Y =M ~O`P<اl.La*N` %SDNnɥbNpmd8'x}:U@5!+5AdX\@~j"鸋bf"iQ荒Oo|zԝ~ٍ)I5J0WxG9f$,1zg3Zp^ Z!z{VF ѓ^iqHꘃ?pַC_1KS|~K7UdMğ:r:7W?uqY/0y""4}O_6Z657*kPO] :&"Z_n~?t8EYN]]|_̯h9 1/Q*w C<Lq5%#2c,^R);2ya1x d^JXUҳ(&S^X"TMԷv41ҘN$FКiFū%#o~R؁ГWoV2([ULi0/ K>e~? $lZ]*5yW~0pre*^"-Ce$=iǛy^(|#F`dA[vVL3L6۽bcλrԻ=HS^SxBBۖ\r'8k%9vuZ~T|}1ugϪ+Eg_jV֤na|Ϡן{\0" ·Qӓ 1{hl"͠IIjwWϵeӪf*Q,ؘTB2:V9Vye λl_e@1Ap#P"V*D$, C+ȁ[hH+CQHRVa S띱Vc {-#c.+%ᐶ1qLT%/3:i7<ٚdcIhT7yDKՐ\_+(%K̂qeεg0foݿ2J9!VR2'llĀ.t9+l3 RY;OH>Wm ;?6nnx|穖a/jjJy@a<"+n ݽ)&--O=ujtzzufPMF_<#&+ӑ[JS*QRgr6 y&KsW4sWAĎ},c  DQFmc*A\hIe^-Px% ebh ^4:܎r07mXP-Xhm 4w&m6*GT KIw0iB#WǺzkG;LQX=$o=eTDAa\*8YT~ʹ;EǸ/gqDlS<`s-: hfba3V1U!sht IUp;oB!gh՞6# Y NVs궪"T.G:3Wv!,F.M0'Pl1=ɲo ;q{׿YM/:Rq!MzhH3!4 (fEy Qz>s,Œ`ٽv-JA=1=N8n2^&xBR!rm|SA*#޽*x7Ho=klJa^iGjd~sVk/瘋Jx->cm^3@F^k6/tC]S4Z0ٶHk?H bkHka+[.r@[0V:mP&ƙARLZ5uYы6ɶG Lt9X !-F/†+/%K,it'`ΞY\9i;>mHr>-'Q4a[ +5<V+bɱEG{BGXrDGG7<-^H"&ڀ޷/Q6QC 1S&g=ʕvD G@9恞ӤLP-itV9̸!t 3MDPR =fFqp[>9M})7j<'oudXDcZ)"V_"D4acFNkm8{~jiS-oy)53&D'|8+߾84U& Q;7 qfA%*!JkBkVcٜKhk!fEJF(kJ6 N9ťoW12#XP1g&6bkerhkM뚡ݦbʋk Z\Z#G@XM=$.ER(I;^t`="x=p.Tn9L0Gn;kHE>ג)E 0}I~a\-4 -BK`E1*Lm+:7lta\ gj8!7ih 5(O+L%JQAQAQihL4y!h!XQ>jdS,eD9,FОkܼ:#syGjiC̛5-hx_}vO$ ' H$y) ?߾@PQWsD}mhn)1H-h[JVJRQ`I:V"5^)ELiY(LAu)p FA,Rg>2E4^JJd 1Y,-,xGI4 4jL |`|ϥ%}k4zV8l49%7Qf"Mݎ]o5qոn>wczBxXX>c+h0Snn`ڄ@ S)J%[sW5X]MdZbm\QDp5:Mڪ]e[\`a?Ü[ =Vk;6PkOtyvbQug&4j)iܱi7ٶ ljp!0j!yxmuɮI$%Te->sݫcz}}DtT\~whl=XnϙU-TzفjjZWMnmz[1mjJʵ4=vÒqaquY45Y'!@'-z.Vnu;]>%,~-(ء/A}yeUCꜥ cc)N SҬ YH)NGyz]~{Y2Rt3-ϲ&WC|EoqUϩemv*4ӌT}Zny?~|@ei)<1 MY96+z"mK%YEE'mj.IP&s FƆTHgj+1Ւ\Lgm#ff$Q$Y-4/Bx%Oz촭% 6X]F*{neje*,oВFr xgxʽ?T. \jaFMߗBCuݯ?eT\7,q`T=-y@h^0/]4<%RZjc&/6,7ϲ2 ,vi43m*#\Z7ۋ+'#_SV"$oݽJy,QR L*m3K|yb\ʴc!0b":p@V,4Xxn,=o~qzxB% mi(%pQ"(<Fכ059&^dQGTS|pƝ+ kobtNK%S A*9*#{8`\V6a cEFX"E&R֪9Lm$nN"yc20- R0v6aCAru9@yTQPS;(S+6)Nݧ1[n7ޚ[ɒG|yC/`1;WQ'ixVV.XNBl!`f ^)u3l`м);g<{E׽^uwU]NoZR]{EW^uS̮{E׽S{E׽^uV{E׽^uWt+]Eu+]gԽb)Nx*4ƽhی뺾 ώپ1ǒ=|ٛyhN6%WBn_irXuF^#thF'CWilE[m=OJWDHL(aBsㅦc!941EO1-yfiڤT" ˜q rB@h"Y!m)rQ9iI" F=I& #( 2\@I 1jMNG4H6F[b42LϮRDkZ, rK.LIUt}m>EϷCΜ;ypN18p@e%b:SփZp9 7n@}DHQ#e~5N$jqQȄQ ]P`Hu5r6( A:MMTzB;X4/F/V:fRLBS *h M|IBBQ/Mΐ%V "}$V(}ݺgvd.>s#ׄgPܑ;8q܁pϖ +rӅ>hL8ESDtο{DuֽK&JWYL;Ũ[3,!M8A)gL )'XtIX- !C!9#'$t;s9ߵF ڈ7{R:*&<LJk.C;jRP2Z^ ј^T{XPWA )4sE,Fl&F-*B\d P Y Vw_}@XY,0况*wyorlcLNS*;iLVCbVX %&R'߹A }4(/U D'DZ$I2Y3;BtH*C[ע=a bo_.{\! L% Tr_iolSG\IzߥY1?ara}8`&Un\;9V*y[>V pH59E%vxAOrh'xW).q`ٽO-IqԢsMs#@ -E+dHߟ̺td)ò*/Ü;?lad-!GU– ɌKٗf+hRAe5P r=CS8e=mƹ|j3BdC;͞Ug7#tVyWxu9=]x n!s?LY[n]r  Ι;×@HL]gqLMӰi4ofYXޣ"8kr0s1fd1g?э2rK63Y'98:'(>}\R/{ď!퍪`XT;(T_8ޟ}W'w'޾?L^4zI[V7Z55WO5[ɼ+7{T}kyg]{k@|0O:<ɇN]~]y^F5?w!JJʟr ш!e_{%qeN:_6!"pe8URQN9h^ c ?i:.ː@u#@'F5h۳c;9_J߶QbDHh85R Մ\JB&p쌺sQP A)E ؈s:CXJ]U;J9*\^;jjA[[J}x`6GQ_GW1`ox+\j=z3q:a%]*,պcIE$ze^wFK }"I9D)$q$d.AX%y%ONт@,&Ebh18ܡu朡 Dy &g G rL[#gCۧZ٢p:* {UΖ W(V;&WktWU ,Kܭubc mvг]xGrY[Y wVU6Ũ>!1E!ZzJ?lky<]16#/=y^״V+`N/#Tꏿ"<\vxPgm}#B%4 %0ρFv#5)#j|&vZ5[KY1Wx7-G-US\|l3+6\7|(lmwB[v?+k$52Xj#K3J1& yb[טg.">ɻxe6CݙѫzBlELm Ƕ/q߭LYz.z\J9(k̭-q^!^(a!^țȻzpMκn٣סqҏQ1yΖ5R{sݙ6.Kuӫ?upyۻ(oL/I8׿ث*øP-{j|Ci].sRdf{_^s.uGNFX7v*3KEťB <S WGE N8S$&CT\:6Bc3m6 CصCWz0:z.|BI <XC#w&@)Qhk0Jъ ni UEJ-tYErM`AU}>3.!!F&kQӧÏaCj,FpˆWF k?xgCKa dŵ-d VJɴ;|g9r2BqhP/Yf i%FEd.!C ι+p!M3 ]O={{=kM'6DB-ME 1AσOZ#U!2^;@A׮ {6euf! -Mdѐe6 IV_?h,lk1bA8q|ޞj-2vvvsюQArA.0SRm4P и@)B Bϴ;ԥwi{V@rxN{ ʘR#P0<*\"Kh7<.scc IIψO93%ት[#g=:B 3NY4Q]VdEXW& 2 3ʌD&32tAZ=7S)E jM\d͆J8젬QgROcւ;VD/\`ta!:/YT;'JuED<DD,5DK+uǨF>EI]Hef+tOoREK5OvhUV;0EtI_WVCY ʰܝJd|Iso!Z橏Pk T:RD6q{~} ޿RW@jB{OF2݋))À=kcF{CS"i-/'X$%HJnܮJeFfZ |LGm7 ,3 [":ڒ N/d5H-k?N'ʜX}ïd2hdaR'ZK(I| yIEt$LEt!Vm%}FY15)f:$(9݈rdmH3>rhAVrt4310BB2PPg($^2!YobA% 9Pg H\558-;J(Ǥ'Ma@-vx..[woD$'OKmqWoeeڗGE4>yLvw&_|O>=FM iH:&f ,cCه8U`tfA V.Zz?MESڸ.촋ǯ f"/T{)}A- ԋ9yر+wWTh c!Fr]<˙rȥBz\ٜ &cb%<@rFϵ,wJÃ&.mysYj>t\KWec%<,q!kgiMÎzR-4!X}80Y[ sZD)3ln%9 ,ӝJ>u},H{Dc*KB*zPDq'TKQ1K>t7p[̅-x>mёvM hTeT=>3"MY('+W_l JPQ0WsY04\`D$_wJPWnE$$ IJy$-G 5(Lu)-Ib*AXH>fhp8%b"XZYh6m%5p6+A>rt샫N۲jP 3Ի~18!vX/ۄZσԳNQ$Àx͑JՐN ΁iɎx[!J{{&%aن C6-vɻB=7q~=+R#/Hs$\Qe8۵8yb+NC8yXz"4h<ódva1pgm6I?jY+_yIK=hT9F@,DDꥦ0+FR_Z=kztַ}JsR?oP+>҇w futAͣxd{G I=@W7c2h`iί`i*wMD6؍稣]ew2;sdYw@W'{&(vL7&n`j@mWb'M@I6fa`:;ْm%ݞ^ttqCrK.ӭ۪šHJ,,Y?ו5%d5,q^*q/vH]?; L6̹8vZ#hܻHl̬>2HU٫~NU6=`ߎxSVʤDU%Ko,Tjb;NҜN@t}2dII(Ft#ޖ'-Fی8`EpqJl4jB8F"az1x u쭤' L&az9z`R{#[URe'Pڲ1VX7TRioj󪃮wd)۩]Rejzڢ-D3AQNH9cZG Q+oCe,iYyxiuË,Ҳ:_Χ/z}5jPƎ6} _Tnw4l k\!ޤ5Rqak/La7|"8v BG>A΅mws>/@nRI^Mx<p[0H:X!rF΍4Hᥔ\)xf%~!_!_~P}7kv+\㌉ #GhK'ȋ{#pa"rĉ#؂q kCTdX/Tj4vCP q8-RLpD 5ES$խ =w>o?]sI_ ze򆤭_KvrC>w5V]%XKȳi5ps}[(p(`l;j'cs@LpBHafcǖ"W0 &,> YhZ 1Vq$\@KeV8e*"ST^Exl4lPMԤto$ÂtU F>â1X`21ǚc"E-9$'T8,X28 C7 p^5t.ܥO{ĚXDkA]b_! *_ {1$$Sv;0O_ \%`x/J^~Y.d/7 U;`R2@û\,`:9k3PS{:BGR(8 3ꥰNX'rI59VָL1ĻXE{=bbDeeD&y1L \;L"㨍k .rIsi̢E*0dtݞz睹s}†ϲةmBX $DJRQ H3@IK1NIĝXZ{nJG*#R.MEǤ"9̰A&#/f˥~S"П/{ x%xOْ1bX1|`)"#t|&c3 ۿ>L ^ΫwJ``VtbR;a9XM`Y4?5_ÁiN<8;MX>t~|fAZN>ld FRo|'zFQ(nj$2nQd=Q*A϶6_.AV)kU𖟲=Urx_ w囷ŋWyl00zgK9/խf0;~d5KJ)}JMːem\fY>!0< N0bA>͚tU ZjɦU(!z$ 59ϟcY c kbXT5 F瀎Oy/Wx׷~x:}ק߾8iq`~>[KwXZ4W/Mۼ4E]Wͺ-7|[=7Q){MTIN1YA )zm>@el7\5E՟bZ@.|BS¿՝ycM1~=e-S0 uI3%je a~)2Tֽ z#(۞Ocu61]Qp0Khb9Q1rD3vs^Cx7üDQ!\EI< +t g1I.J껵6wR}xH(y$lz'v]bQ..#?gA VłJ,+c1NoS]G*Y.ɥ֠H#.h"!\r"nE:"X B*Caj:Vc@Q{-#c+%e5p}J|0`M,Lqjt/JYwR ﷘pCGѳb3Uk솩+)ѴjH7Yr-he]>2I\]I[nF[lͤ4o5-xӮ;[3 n}R=0Wj-p~89on8aboEAVuyQs^㨜fzid*$yLGn+,y˭=mNƝ~^dyᾉ_-\Fu`c3XnX"(#vJmw}{Lo 9 sIО)F$fLUb)1H%v*0JkF,B0X`<)V8hA #(H8HlΖo~ $Θfs6xss9=/l}_V]ηFsly*iAFjD!5;g#`4SvK;`A4BkM̨׷ۨQ1,%! jCҎmWYπtf,}ؖ㆞Y 6EiJ6e'˒2Tf̶jY&u*30,Lo^6JnP+mCSJm΀y6?][s#7+lRm&JNLξcȒ-s;ݒ,ے-[mKUqVWZ&Ho=ĝCݤK9'9 |Dlb,}2 ୍a1@ȨU#vnR׮dqi8]b_XxzaYbq&d~!:׻mbrmחUS͟([s+,+w%5e"[cޖRAҠ @5ۈh9̞]H+kTQ@z- |@uzc?lCqRA-A*R]FJA+XOO+GBUMHMZYۦZ0M+̆IN/?Iz(_ wQ ZJ+YW.P+ a*WOp.USp7b7p|$v_7~Fe"c1p@?4훻^opכGBW9'Og kMKCl^`47 Kc* eM6kl۝OG0pΧ%n=r+OI%|JvΧgΧz?SlR\q36 X] DYc΋eu`l$*=P?uz~@f#l6Δ!Wū1In78ٰ--e._@v1\w_j<KVF} 4e2q_!ӥBUR$bA#*=b,AbFf."FE*67׊^-|çO^T|#PR$8lT$8q*<Aʩ3)Z7(:Xy+M{\?rGvog ʩbd,{%kuMt>FoٴId2ɋUUp Pcp΋t}nvύ6j4%GP0pPs`SV&ÄT$70rpX16Zd ]]ϤZn R`\'G^=i nMA-AL (ؠׯ{4-?P'flJA bLcE2+uh L{.:מ'>.poDf!F9D뇧Y?ҹή]{`{2|| MVX=~xam%Wp~Y{T5yw~2z dU#39(e@Y1R`V9$Q@J(tNZڭX\RdHf_19xHIoJ"sU0 mWkU4}թWe]gR1~9r-;u}'B4&8 Mmkz{تU[k+Fc޳R]N#\j&H*"1gp&E͵t+pGhc.Ky]kjוՓ#[=\8Lk.g~MeJ`EBb6Ad m`XC,l\s;;ЭHòlZL92RX ZRB|*@<аN:{F:f3 @M $Y."BIs#@pV-%б!N:M415m5b^_`DUEz֛ۘMһܻd-%[X G} ˙ޗ>Z!vORv}^ba!z]V>˼KVO{oף@"x@jlvMG{Azj)Z!X뽱'-Q-Z떖\YڗciI:{d Z*sUEusU\ :sr`IјOxr@*w53jjo} n&LcH#&{eFIH\~W8;RHƭ4F"[1>Bڟ¼\LjQ&*5c~fZ3eA쑹*AfᾘByTYK4Wo8iuWPe|߇óϿp4UgP{jF{~z>}Gzu 8b1 BRV(Ut mdjwÝ?~WGI\4Pڑ9U-s$q*uA bXĄdTmĂ{e"N08>YY,qfޣ@H!XO6@!I6 jq*myUva\5\- 4<\qjdz\<(fDdeUh!İ8㕮GJr*OvS`luٶU e2!Feń^@H#̤wJrs^3lԢՊ5epr4uQ;cx"SL}騤DD+ִF%'Rl? _ r%qr]X %'4U=/F!(0}CtURꉑ5yIQ(!"i)J$07.GTR6")`VuiH2ܥCң%:l:$54s,B˺9k` A:MMq\cAzEp.pn)\0%Sd(dZ'5=rUw*nQ`1u,Y! A.IHzF!LF MF0Ac;cJ)]E~[^p*UN NoԸZ28@wvRrgQ?[ EaɃȢR ʻ+r&99_{=bjha8Uĺ>{c~ (Be2̑ۚ!Z CeJ+ `C 0M`,v;s9[ -WL&}KR1dZCb0 cTelt.a*c`%(KzY!7%sC  \BW)nbɑ"$-UZ>q&4JurUΪ޶q*K1fLL ؜4͢ҖR039v !M $U"DF  ,"˖4KJQd_>eJNUCimCJzC??{!9 B X{P0ys:d|Hsr%$4]X>Ta|B`ƽ4? ԱVMoӜ/%O2x׻L׸GrJp θd|0ZJ.Áө=OfG͒]* yMY-'ɍ_!'"EiP5s͹Ê 9\mH\"8y%:::.^8?>k}=٫Iih%(_Zq\꫏_B&eۼ]y{KYxz"8[BUR,NݗA|8nx+%oE$j?.!|xusG`lH9t$PG:]5X=LtEayC`X(`bYɇǗcLny9p娂>%W]A a磎Km4ӑ2"pʣ/drD 7Xy([I&lv2hn1_ϯ^s{.zןh q`- +IIFX7ڴ547Z"=dͷ2|qW^15wzZh $_~'fov)Nfs* 78 Mj~:S4k[ST5TX8`GÀqs|"fobS̡$ˇW=R$!)<6LY3͞Ɪ_UWW4?ݸgUw#6JiM rɧ%[ 8%[T,9#X9*-)n9a-zfVp•1Tcw!IJ͍0E.D|0arroDfEz jw-bSHmln|'>1S6 /GkQ |LA*Gm7 ,3A~6䮹tD-=_NTʇA%'ܠ0-G dU420} J\B:`^el.T1 X0[9jWa iU3ؠL1hL>% е!A=rɃ4&=q<+10BB23u2?e$&F:łKsg Ԥ!)$5[Nk9}rڋ=0E?(R`K  ,-@C%N6!lihLkBH#A`GQMp9OuDC((9ʍD"6rבaYA[R.k%#RH.E)tJ&vBCZ9𢄕yA%+J)r%B#A8a{>s/|nB䌂Z($RKC K)Net) qKRQ`IJP+" _3T+] )S(oo,Y"R )RR%nbi1g;Jaظ (9A.L[JdbHa ] SPm~I냗;#_@`PyFe 0,UiY% jHI4 1N2[W%M!xɚ08U3b眼Kmkɾz! g=kY\D&S`P^0yAҎ(ȵV̞C O:$JQP tq1FC`i$"JP@Qrgr.$+)+)EJxpﯥlk4,1| >Zy .#?5ƪw^-3M LO #-sBQe8ۖQ^SL q7LqnfO./U@ GWӷv1aLJx^$/voE/]}i|z]ڋ9 [lK*=Ҝ6-J/mRG>~03囖oNoW[YdN~`] 8]_ɀRw ykA; ݢ~a6'[<_V%2f˘c˺:?2cLq`xt Z}nqV'|EսC~6d w5-יu=TB$Uk&۴3h7c޺RgLgVa^va|ݙwPD3AQAHcZGQ Q+oC{(h?CA=ي'\F;x cBL9 AxLZ#<LK3:=VJQL>1waؒtP[ rR0#4'(xAcf;9OVrΝfb;Όgޟ:{{ѧ `qm~W!6^mcӚG2TdǫFKL ?k1vND4*k ƭ*qXc0hUA:St[ҷGHRgquJ^` f;k3j0Bۘ"9H\DGmd^,UpP&) NҘET`ZYv*9H~y$LViǬQ{>엠߭Ggv6r6w*+; B#L9AgRP;3em> =]&?@0XYi JwrZN. ?5%?tE _^UD>tv20>};:7$,̾VWIo3@JۜZ#'/4tht0xRasgnLuθyr3¯;?ߎ gb4׏|c{y5[vQKQU;O'B O$Ɨ?116>fv|F?d`ŤdgvH5>O Mce=}jv4„G҃9, ];#~pg0x"RTǸn|7W|O?O?_?^H+~;`À$y[觤"e-߯zIqHl>$  gjz^?`&̭"Em'4AG>2ףyͣbMo\[D}> hԷ D?z~|5vi}C(.WҢӃ׬BZ2O@lPt5T(A*wKClNqO"EceKsE̖T1Q:M1lV0n `sFf$PݫG!=Rr&Gf|͟@uu6@FV+-7a"g3*4F>`QQTtA#}.JUqfe(j <=gQ5& ]TY)q{0/[ʲuR͉۫W۫{ce)ĩVVK' ÷qou}ۇK~d`X d 6^;Bqxtt'ژ.hq:n$ZGj+G^ "6t, \CCE4mL`1#QHYu2L-Dgj2b=6MX!-1b6q6}*Z &2W5(u__ ?nZKmpMVY2jmgyP5L{֕ڪ\ B+nLB닻|Q7 bx5EuumkG5֎mW ;?\zas+Wχp~yfIO؛p]1fۑ|[HTNfͥ&DlZn=o6EgmR>ubsBcs.׶k?_8^[۷Q؝("v f+l%"2h>bw}B>=&BNhFg0SBUn)1H%v*Ji< i),L]536CFh;_e%q?-2Fsly*iAFjD!5;g#hG1<7Y5؁"h@Q2)oQ9b0XJBԀRELf &z ^^=t瞭~qMʖ %.S#15*Cz Z{1[,!3V'cm:Sl4oqzJ3n6|e&p]r#PB}Eo۠e[k}Y!i0h6w{'&6ͿWt~껵%hl;i ~miM lYFCÍၱ+e6(SL )&mT/jxa)!ʙA^(, Mb1P! P8,0yIo{9GgNT n4GjņfԞ-R-$0VXOFu"9BZءUaI):71CJݨUKBq 1 E\!*+҅Vpb#VnI=X>畞C&$*8ʝ7\ Ø8P=t:FY^1Tk%bc.m&{Iv(~x7"f:_̝?W=loA#A=BTUUzui0?Wn"=I߂|. [r6;Kޝ}#Cq~ nlߗ_>9ߎnjr;Iޯ MJ2S+1L@/J Yz xIԳO͉3U❳;FpDEp ܗ[hʷT<?Dm{}~W:^uOm|lh0[x@_zf۪Q[/̝=oz0{Tԛ*5Ac~DMi\0sn9KR0|%gSn/#0r4ѣ"ݩ+?fWEG qu3Im\>K?'Ü`K O+Lwj%?Ɣ !LkA`GQMp251RFsaYsGEmZwՑaYA[R.k% BKl?,ZS?lpVx1yA1(GG<,w9Be>ңWpPH,S=Q4H%J*# , Ij%X[T[kƵQʪ脐R8ÒE*,5;اiL.DpCLDK9 Q AY&wE?w-wEz YûAy|ߑG_ ?l)؄Zσg4"h-H8aJ5G*I@WCu Ai)[Bܮߦdל/tPQK.]|sFp*}SQfI 0LACy0EJJ;r#E1]cɮH6D{$T(0cL04HD"3REΒ ]YҞ"eWvpϯcU?@_4| >:y$ab}ˍP3+길_G$TQ,)`LT> l AڮBm-Kj77@0v37STD:w3,7MgS{Q,&COtn?19-wo2O#|dU0SM>c/L\Mo] b)gdt2!(QGn -,Q#o> ̂HB9U R˽)M.q֫h|rGɯFͮ[ss dn{Z9`-z!*$[yCdt_2[J>Whv[)fj,<>\pY5D-.9){&GHy Y-P2zE0PT,S\-&̀D j0D0R-c6q6Xf[B2B½©D&)B N {Ǯö| Ɠ`,(HB3 JEB:`^eH ) Ӂi]I#&[H</ аȅ4*C0 A|@)SM@ ေ>rDPAC/;tV!!@: ĤH[XA`Ir8aKe$sX˂@iuOM{VuZ~/&4Qkgp\Z}Js,: gÔ6hEv_s`8LPT|X`{\Ѻ\y:~0yǓ$S]o5b{"pl%K~ڵq2jF;x#9# 1 <[pr^GC*-Zeqc*h% 3@0 ^ИyY%84W d49iPÿӫkZ64ٔ9$/_;vWy@n@LWW1F73偛7_R#9ZHi9g&ïOyI:/ËdKY㊛}f}&7FA/n&X`%qG2Y+냉KMM ` Xy${^^:L]]*\|:Hst lov鿗^\C>ŐST[M7'7{NA~L]xWsg2 tFhIշ :Hg-!ߝf.N1;\:^o G>}pɬGO?8+|ȝU;G> f]CP/RQQ\uϻI~mûI)f3 E)@$KOw#\YęBx|"T3@a |`m0%s4t~m_AO IM}}\Pijk^o{CߕPWm}Sy@󡚳<`\߷PvdR#\[:ؙ<4]uw?MEM]=.zٿZP&p@g7$\$\t2۠t/QD#L9]+`ɝ+HgZ82ЗK v2)9~3CjD%),3ӜUuuUjCX 8ѠFN/ާ첨7{Xƽjz>Fi:OV 1PFyg6{B(kB)0ZR3JAR鄡"]n cΛl*ɤLZQ5g^M/M^&Qig6^&¸I|' dWOgEWB# ~*p/#F;\̛ 5z3(_Sڍvgɲ nāS~ݬellkBiWڬZJ;ĺAW6}|H5a,ꋣ!u!R vBP,@1 WbZUTB(FBt5#X-b?b/u F[Ke$`/RKJ1Zd3ZLnw,%},3 ZĵVlvTlV][\zC[ -hmo$VnxTE؆4&y*b 5(dj2lYpfo|P30{ߓKN/GopR70}X7ĸĨ?\s!u9 [C݋x;}aqЍ u ,200_ ELX߁;$ fK{ k.X{wh5Z Qר2 Wo7 nwX l7:}eyϋ/6]ٍco>KuJm.K.ɩXrZݒKTrk4J19S",ٹ|Fn\kғK.dٯB6t:[c(^!fH Y@SJx->c}r e.T.J^8;?}֥ P`_>: HC(&҂sE.}6 >_K%&\TYl%=-3/͗w6wbmnY[{8VzݖH\VW]&"adHyBND0'dH q*NDTdHT*9^1 :*8PreQjWr~59Cix`,I鈗6(SșARL:pk{oێ挛|[6 U9!ʙA^(A#Mg(P;%Wp兲a1Y*;09ZSH;蟵3F]`tK.-)ؔܧ{y5o5u]On\MYGuf5fl׶n1w cs %meCI/ɅN{gH87A0ɵ r B($0f4j|sWL#MOy9̗C5_JpPy^t]L}c7FxADu1{Rɮn(-v?6@Lj.Mf9S5qD?Lw&M5Z nxs7|3\6ODg3Vnҿ_ܺ99!D!e!x0k5f,`ZF hFs+@wuIWi؋Q;8p* X8VV.Z d%*(pe =`(bB <* 0r!,ȍ$S{ `NgY/jTӖ;Jݕio=:d:H&ao*i`Q{I܁'' ~yKU񔵗TyL+YR8)'Z흗p.gkCcDTAGvIu2C #$$C(pFBb%#1 )DXҠ0I̥ 2N:IۧA(fw ITZ$CV6?dۥK`AJt`%EbH4αK b^FXzKa+6>` rt!kCKֲ%kYv@J eDIb6}0*Y?dP sZD)Z2li%%.:wj%+x ~u^id4,(  N22c#|{*9ʍУGsn]Ȱ,^ -J)O0T69]RZ蘲N2i1pdjE:1s؎h`&A@ʽAriyB䌂Z*$TɂĔ#"oAs߁.aS@I*# , I`S'J$+9 BGXh)iXHEߗ!"}ʥX{)S*nbi1g;Jaظ (9AQ 2y lSiyParS.9~̘|ʟ/mBY'ڧHA$x͑JՐ AiNO ۠M[cI0xɚ2x.}]o;@ OvՂaOQfq PLCy0erJ;r#Y1{㡎<܂gˏv(Zxh/dj@}fbfT3cfTr YܠТ-&2vOcy:i2n>4Qֵ: 8l\ l]CࣃT?z Ubj_Uk?]gR#@GZ2hi#q+ZcOL>*)JFָf4Gӏ4}FXoqc*n/nwK <"BXYL^jʈhA #(H8HgK7^'}i>]Ŭ#CV'+ػHs1-`ntY'7)`od^@7KSʟǀ܇ ,}nI$Zbg%۲y';|w' $N0'su'K 3 ŽcGpcZқ^9#I[jnܳ i@d6K24z xQctZܜ t6Vݛozvu+{U+_Œ! /56G{EvR>,ql0^͇m7)܌z[g6 LR qTm5ֳ,nA6_LQ)MT2m}svtX}ɼ*6DDtҡmWfC}bss5N''|@.+bs ܌;6PzgQ1B 33U1d@ۛcÑ7$\^tꨓ Y )qujP{FVPƁ[o/ 8gpÑO+>'b`jΪ?߂6rgugSуG?n'AC?ʡ9ǿinT t=$L@.*@Yꥀp%gզ-Tguٛ&T1 AЂ6J»E4}m='y6-oA>E.lAnh.o+(QlX%v0]ۢA܎Z^BŻ/jSvd҈t3Upcq7HMia@/fϳ?wXzI)~3B0z设Lٿq)dJ3AQIH%cZGQ Q+oCw)h?ڥKA{=ݝي7\F;x cBL9 AxLZ#1waؒtP[ rR0#4'(xAcf;9OVzo[(rr}3[yOm^d)8ƫ??uszgμ>yS΅/bG-$b`[ N)Ҍcp`"YF^&yȃL$8y8k:.MO[-ɶZ- buU"Y0HRk4M|ABჼ2ujUТXőp.uXE0L[*GQ2:IKc5_@E||wcXЄ4Xs Ԉ@Bs .LzJ(tmW=[^poBk~uV g J)zpvS~m(I59Vָ`o{g$-]OY B}{ŏŔɵÄ(2[#`(.Swsi̢E*0dt6\`g&4<5fi+X|_T;vaJ*8^Jqp8V9ޘZP:7TxG$t!:& 2'TSi%L 8P (O ' XƁFRnR?, (Aai%/+Fcv{ɊIjG%h$F+X:NC#i``RWc_1K٠<}ǎNYѩ/g;q˧>Oˇ'gߎN> p`, Z$nA8|eSCxTCSᛌ˚r͸KԇٸZ(G};.fOtLg j+6vǙvQPms\&D0G..>+GY+1Mo~zznlɝ1fStaX9oWO𳺍ב(@*0wHs=ySŧ`P_6!|EXic+rF!l=Bc625=}tyxg]2KI55!e! ,J⹆ )`u@qcJ k7لکlKă-27mي9N`'=;xul[=y+d]]_8a%Q,ؘTB2:V9f yjdz\ R@1].Ap#xRq "eĒ`IryK[@íhG(:oЙ3ZDk1hX!- !)me yp*dt];j{R1[Iu|IX_&/2ȯ b}D$evMUxI4RMȕ_u$Gh}j2ோ[]sJZe 9lV[wJb{y&T=/\-`Erq+ Xc7*KD-n+1xU"X1W\\%j{W@%Cw)G X썹J䒽1WǮJT2њhZ=2W ޘD.bJTr֚h ŃTixa'e)_@COח_`R6Kx9! # [;ys(Q{drDbZﺙ*R~fZP&eU"XO\%joo/\I̵Tӌ'IPQ~9˪,?-޾RQ,Ƀ6gLTb= l AZ*[8 uLD=S{naAFlF $GC$cʑpЃih ߉ryK fL(s +Ɨa.klsmD*5+11W_J#yׅwfbo-d9F_zLj[F cOs#͙s4\Qe8+Ԣ9 Q^Q": aJf 0H ҕ#D-!2pM-۰rY[N`E.bǑt0B (>)Zsڔ9vӕsDXRxD% x/GDD ǀ-xF)Da䐤XHMLc1>ib`7})r#r]w_z%yFTbgSɩI ,b\)JaWΐӨ͊\i%U[MM1X8}ĎwK"L0(<RtFc'dA9zՃoMH`3.@$D0 .\*Y+eXc;cPQֵ||ÌKY6䞷d"8Y9yADstθU5.&1_s=bfhv=S3ξw'UXᦲr:HU ~}:Oc~ (H&EDK\Fqɑ[9G,ZCJo;vVTiG}B7qG=9O 4PR1RjoSq8V9ޘ-,EG8QEQ*[Y7`6HD+1:4:o `AJÍ`LrtGߊݵ0(ua/SP%) #`+5h1M-uXls6H)-;QZpaqT]l`gϝWWK_Z$R9l8Ib$8"QRи?&O`RnR?LJ`0rCP(`bZ^wbrQ Z=j5ɺQ(!zEz`RWe? ʼwXT Ϊ ˙/@ӿ'Gg&㳓G C5 nE[pg??ch0^54Ul&g=e&&\3)cNo0B(=>_8wq7[U.vҦ|TMNW2}+@6yov[M(6UL9U x#P~v6 đE#&y9;@ o^Nn >[rAF&cX>`{y[5͟y# +b`rBS1< b4ުnmT;^QHvrɥH#2D%3ȁf6_LHRVa S ZG-"﵌FMFSpd}ܪվ5uAe{SN^ _ - :%8YrȘ6QUUWw]`߸pR \eD$mfA -WC*X_tun>Md~ЏʯxZ9p^uzEr&Dw mmrxU[9ʙrM:2[͟ SMjo+&rڗĊZv=bAn+Ċ#F%./|sF.w%mH(: KHTDj0VIeIZ'E$7.K,L0D^$lȂ\Fx0W;sW#ѝѝ\]ѸMZ]iZ:GTWt9Q *"t\/*YG!dM}᮱9[GF篏n$JaFiO&hvzu1 > [[\1R:빷NG*b ޥ '4MmK; &;!s^c&| C,B(d/RG<1Uܷ.؊g}>Ʋ*黋 6ڮa'r e[YMIwOK|'!gDj:i7y~&'[TRќ)T$KvFrkcu155bY,=HO?Oћ>tz_1Vb GX{(D:2^Id8&"Ă&9MYzi+߮א_$ e>7?y !^zm_i\URd~zӴ)H?ZA@q$c+TiN@1bwzFh]Ao[4xCtIDBy(s9v@/]Tm% &5/&skA!HMMJkJ草C d($u< ""C[Zciq@990rJk'Ft2(&B9=rxA 1.D1J>'>9y)K?eCE*Cnj FXP3#jeW%[A]V1@U [IJ]`.4>"@;Biec21% Ht%kF{ !d |l LȦ?(`uC[) $_{YhoG%ǻ6Q!h.iZIkњ]cW:}7Ʉf^~lHwXV ɻv͉\P:"Vb*UrJuZsTGT*]OXIhIDvdؤ{M"O C4!4~k.J4}st/gBI Z0ʑJkɮ-8-RYN% XU)m\IhBBzt5H߈U{tC#ە:Bv:S-,Y(wޤM> P'] p|PqnDFf ] IwyL![LW>.11hXx>]+xG&#p~uu[ev$8|%s9 = VFzt-Q aJ(EQYL[r"d:ɝ S.H-R2!XB$s3[WiEf7I!P2^5><^M_B7?62sJ -5,)-.ڢ"`Yh&o$Mի$1 !4mIkJ-6!Gm*.sREF ֒9%zrYd) %o, ]' * _>L/:5*k:^oػ?.V Ơ,  ^,N _SJ+Kzk%,QV 2T'UR*6@0t28"e"\x:^ӊ; i<tԞ=]qs!)I'@IC&-("ؘ<Bsj]DEJsMa@2!CdEG-ɚy MH#(Tt⑌l}^L,'F< sǾQ4Dl! ΍2[%\1VdA*B !lUc$#WUvX[xW.UqZigia` 󷉈豸Py7Aq=7udI$s' l^doG Ǭ DRĨX(J1T:@B7gs|ka笤Nupz*_ٛ:oU@ը1g|ԯȧ% J$hAE )B) G&n"G wr^E݅׽zwMz iuz'{K1.ۃ;?fAu|Yۏt{ ~fEHXwaD8,n>ɄVdE#?fV gE-8Ey:<_NwǙK8c̎1Oǘ,CWc#3ųc7gbZq^-NUIwa34D6Ҷ-d\٠/hq'IʽrwnΊYzZN&R+A+::7󘩏fdaӰh3_9L~x0c:s6$R9<+h6jސ40ckVhzLpYcZ:pN[+-IE_UO+jߝàQq86r4~<3\dқAGyЦSI灛*A* 9wBt '0\Ikv6-3gRf-l:{Yu9,{6K?QҶ8$oFiY^,{ҁQI+lG1e&~Ce;P2^ `miERuV( rMJ Tka1uK[a. dBDNhwqU\`wӈ++ qU 6bU%KWDՄJ׉O(*JqEڣ%*J%tS+Gyuq|6υp0ҝHǸf^c**IP:~~~$$$BKDc"-Jz1!7\-pn4 zH&?_Sg4 =!/|'˹+JQ`*h];y &rOj) ~>#&#BMS'co2vh;p6aYg:ڱ}GX7h|GO+̴6` {=m=ռ7&'Mr ⑨ տ'_Xe#!T5Kc3,Gb2W}_.jL;tpH];`3g#tbȧUu3T`aΧ_R];+TU:i]Z_dxbj`Qzߡq Vg- nMV!¼@dsf]aKO#vTV=-o%d^eՖ (,j^+TRuB"HL[ %JD#H}Nd! O"TI—Nc4o%4#g˚߳N=s3aﻨ͛jm&ϬfjvãgZxGLZҦ*u׼ Ac)&kl:a%\756c-'?nQl䅑p0X+7g~F/s N> ¬'6C?e-EN`w'ŝ,k]ZwKxkK|LvBC? _}>p?㰁m C$]$l[.Ugھ}xv;#ݝNfȣqg*2:&SZٺPDQ)$c?~Y'&g~]~'Zj-4;mNFöCV%Y׳m fG|-v&JAkt% so TXK1O\iN'~!&rp$9A4 MvBYLYQb^X+FyboKqjw_r}M15%Wey"lsU6{4񽖥Ho2qA)&YkK-]"T3=B{簗)hX]4 IdJJp(\/dSBOk—FAȭ"kUp"&Kki9G0Z&>')ZVlcDzw:t\ap! RrG &LSV!E}bM~gzH_!}ƻ9^bAW1E$%[>T )␒8)y X8隞|U]]u$\ŀnC!p%ǀ1!EZpMsYjG&B&RbA[Ah+vd=*:4SNq1ʹJh' +%=Pe[踨|\Q1X4E.jKpFned1G:xL"3TdB'cG$cs?cE~/[־p m 6  * F'h/9t?m)A"Bb NT3MڣoV}VW?=] K Wʛ޿W<\L 5iv)E)O?a71% |"Xt[c'aPx{F 䈎@Sv ÿO6lx-KwWh|J8<0) 0(KAyPkVDx-rDcz}c[eL% =A*~ R]yUv#!J.( {RgUZXEg%29IOɫ1)J|BhŬ@ JL ,бAؿ [iP^F8"2%NJI AdP8g<&;Bt'H cZZ2!ߌw6 q&ʲUP SraioSG$?:ea\Z&`&T^ ̮pTIޏIJג [q&sk aPN%99SsLܜL/NٿZh9+yEG)sF:^zVE\?Lf.ê&W科 1ԕy1.mHU•};_޼WTP`Y ~5 ݣqВm?]8GǧP\2tѽjx3ewWwl 1.Ssai?;֖۵]! ʊ|x;÷@HLjv&gh5Oy41Qy% 7ٿO7<"qVF6:{ȦY8cf>$=ǙG㲐WkzB{*17lM%?_#;>?;׏?w{wwhu4x{Hޙv"3SfSs ۧ愴g^h]#>Urҡm-?z?_}nOwAwӭլz_Ƽdf_AQ/慐}}RFR/BB4bpYm1 [16HhbT~[(;OO1eJ*|r5rW1Y6>x1AμG# [(>v@ЇR^ZA1 ]d]@pBҜY4u Xx J$B$Qg)q6hp?Bo-rUJIIghĥTL$АsA3CMP3@p&e6GC7)Y@oFv:q_L|> xBQp,rV%/qa\"9佦ȋiiJ=v DSB"i$0 eF1.n:RD]Evg2M(:hWmٞX'* v/>oz g+X6g!ONˣpn(-k[ Ɍ6w^.nj;9I#p&Bi˵:b ɄlQ@ZՌRBrCRrek 8l :%H[J؄Jm95fr]gr׶k"gvaXp.EkGuZC qd ""}y⩰:QB$bbZP.ᑉVaRY2"C E{Iޚ`K1w#օ$TGZmg}9ab<0KnXzvޒ= 9+eRΚd87&zv҂DN:Xuj6zHP^ t;ޖ'-F<*~Fޣp EÓaLPB+%x61iu3jQ@Лӫc9o}¦UG}qSwSTä+BnRQe'MtbpuF̀`, BaL*Y-.vYɲ_mem(Zo$m YkJO4Bǣ ADUP4 bMs9Sb(SLԱ|xǸO` yjĶF}J|DM[nzq9c ϳXkyɦSCO|h5A2m41^=^9^7ObV,VAoя!x_;U-%/{;pp)b4ٍ`IN/#W Gonj/nQiح(%ep,shPVx9%2{‡Z8tіJz2pl[sV<_5ۢ hJPkz2իnkYh6EͮGx8{'7# LQyD^Glb|C#Q&%QY}dlQ%:QYEFEJW].K՗G=blι]q立kdftv}F[7?;u{^};n$W/uufR4+̒GLP$րl&eQ/\,wWJ2"dg{O6D#}bB tN)iYeʒqLI3ER@eiҸYN2#Mg`*I@8d\LhmhF~^ii$+ܝ.IDF^´&9;߮!찔S[CX֜IhCR)IdL!b"0ňE(yURZm,TQJDQ(c d 5:SN(3C2XʶRM=&7!Hg)c)r<Ȝ'UP(J f׋!6R+ޯi׊'[!%5)q["%8g\2qtƎx'< ˣ99#fvn =fD|wtPĚY].iB]2:<~Wmum S{/{gu$Y}=~'z:PA30J>Q*r~u&|hcpAk -!8~wo̟k47m3'~8?x 㔻RS?̃tnۥ@ȇoNp16ɵ#Z?aa,oG,XVpp9՘3nUrӨn1=Fq,ʽɏxo4MMrSM:^쟝:޾͏T~û/~;.7{/?Ъs0m[K'!~-6CK |ل3.| 1PqgEkkk@Rc2ݛw(hwʦժ9e;S/&3?OzqvVJ\IOSB,aMqnܭ,e#]Gkg8ՌrqVN{ ;s0z`7'ݦ2 ,{ K9Ro ֏TI9ۤ.  4VDK ʻȹBE[Cg 9b[;RёnZSr!#ĀB,v)=d#r<6VW"t4H KMJ^ۄ uPĝ5-+k{b}GC/c.iU~?dޝwQI|%.NRӍ)7Aτ)LWgMzޟ nsْ~,:h"-~.D;S2K\˓ڝT!Kk9 B9d,K$SNNn(`í{n20fX?Np}] Rާ麲@~!a @z?MzAՅ~ӫGM}] d."<^<:q~צ+yAJWm_K`m65 ŶCZO=ZN\=(R2^!@&USM@a B#}qSV]W{?W" 'Ok=2,'~pR!8\d3K" >w*hm庲z|8!|$C%ܒyN3J)r.&aV+Xj`ƉRk2"lIx[B/"D+JfcϫkDxfJ`?mzvy(Eɝ^P/x><_uy7T%gNOVo9;N%x0U{&_KJ[U{'$c$3R,A{UF%}ˮ'&3Wr##zFmMN)}ڔK 6=M<&3xHs-= |&'U3Vgg܎RVӌ;B*B ʳ_X4v~w$J F3RȘք~-f'/<laCg8`FDŽAg_ 9F,BN2Y+냉KM-a$EV 1.S,I '˧۱㾔4'և|\dDv,z*O>>{12̜,k団g KE[Yfu:eiLi8n^ɀRu5'kAT1j(y˄sk%-g̖1[<c.i-ŁefaGkyqcY^Ml;i+DN[jӰc7@dCm2.tUq"0n1I%x+Vm{MϗDտU}~#:zKNUmGk(қkbц:<_O' `0|XlY|f Jm^AV5&QLTy}&m]󰦡v\5XTi*m։&IE_ 7YրՠQ{|ې+㎹p07# <٠iԄpE:)9GcLkmjsL5c*@A=nlmgQIfa˪Ƹ)$Ǯ׭a\5B7߷ o:LL޽w /{ʼnfs˵"7AFWކy#^6b䌾-'JQv@r AxLZ@%%bW|ΊW)/0xsD0E ה~F.V/8iRN=J6S>/ @Yj%1/%>+vrx?2;,wŝ;xvw nVEc}ҜQtX\jw v~]f WVV~Yk4i`2QY{R^Bٛ_~ʣI=R ܹx<՘=SR'ì+}KS϶zw).v"X#u`qWQKanTG_8qvWWN=Ɗ }B J䊓WZ-]\V\BqEVJ+}2 } 0ؐJRz*Q[q%LtB*,NG\%r?qEcWʹ5ۊ# VtlW\OE\%jѣDjիWJ=[sǙlA|@ws; -NAvFI'Ɂ,/9D4/?~]6ˢF 3eO.Υ٧{wSN R&h ct*2:QбDP~2ZP"(?!qk)OF\%r<q%(EJTbيW($土q(,NHN䲓AWZqJ٢W)PMSOOs~7߿SK?gW[>\SS[36hLt!ࢀ_$ ;k"F"@ZqWVgJ!g1!8ߞeſW3>fRr$;E9U3%&TC${`dg\?\P۫3Β<8nsƴʍ!GH` ҒlZ~R*H(鮪D&[i& ^Truw0&VǬRRz)#3tS8G¥& kEJF'5h zqy04FΖCSS FyACf恘(8 -1D@̰2)6 CPzhd* p* -S-JC6xI;פt%|#{U0ccXe201ǚc" *3J&( LzJ0GXk8(T RRt5Ơ6VȺX!lX!;%4d^0.SΓ^!b͠|5' ^=ѠEstθU5.Yc0{FYq˫6yA ?xhL \;L"㨍k .r8l Ah Y+B55gTVxccₚiY滓@KŸrD@ 8PR1RjoSq8V9HXͥt0˺TxG&I™Q* l贒IE8` (O X ^] C)Rx^@Y@tjmV Dct2t f `AJÍ`LrtG.,}l&h^:J:AH) E E)ahSaKP22ROp!} |*knv ~sPH w4I>6b?'Mgυta@dh %f ^&(WeE?_2yQ>dKd=S5; dD#L9)(ԝ(h Bm0[ C{mgTO2'ook B:3='@64(2ULU"DمH';e#:ZiM얖?!-2fKp*ǘaf4 ɩ^@q xƾ@WQC{ۓokC3M'׭bMXQpPKhbQ[ϨarDU^:2y>F{5KI5kB J`ѳ(&S ^X{!QmlMWvDP^ҟrux:=uz[Āa|4g]whNn7CIo*dV׺~ |j>|Wh(LbmvC.}FvHw[Qʛ[|`Z_u ͡tyYz}0цJe6u,*9̫պ}Qͳ›+nlQ\aﯩovnm|3 ΢mn؛pWzM3CtiN?)AtuP&_,ɫG}M$5J:$ }^sku)$ yܸ[/Wy؀e^Y;96GAݟMwG(trBG4ǵ`[ Aj5Rr `K ,DD""&ZH0<)c" >aנb0|x9ͮE|nǷe94l nqg- UJ1LBkw G#XW1<7Ʃ94@"h@Q2ɮoQ9b0XJBSELq"g3Y_'=9Lߝ,nQmW*YW4|MiIoP>H犃:M f LqY̙C:<N,2+C`R0C# (QvaLbKOYOl ZGKHHx3jAx:FhQiP19ZI|xzo Sʻl/a.`/ Yv} l Ҧ dz,[}_7x%W׮^LJl~OZNrz֗yD5dF ENf,J [ lBA=J|`;nt1] %7a4]ZKRjs"D9AGcw_[\HӯN5دSz[cR;ĬW逰 sٻ6v+< ]kfΩٵkN8>QXrSN$92eN?ı[-..B) Ur9 kL<>Do۽w0]O&ӓc]k:d;2c1-9u}}>]_t}k YJ:\`xVv#8 @ AbF6cppv,}QWd~ IY2$7SؤA\G);b.Ϳ8XTvdsv_SJ=02}?([WېUnMwVwrwr#vNc0w`}p[B&Mu=xgHf $~4PHU)X?Ɵhm5BoȃۚZ!-^cjg=-֙I\.PS0PpJqPDZ|1rZ)Fth'TXQr&TexcIXɘ{ENG/svr>/p*d)wR]? {C5%M%=!T\<"kЗl\q !@%wEFG\U+tD%Q"69LBv6jTodfndR?6iqb/܅_\0bvz]2tOl=窄~XBV檬w Idd+ch_FYfcMM|r7!*អP "Tjk{_9.Zpy*;NQ{d<> \͘3Ї"[RIc}ʱxΥBmQ[ vÊr{Y-^&ն %'ı*rU[E9pԷ&}Ӌ\0 "v}QwFD="∈[#YN}2E6g CR)87&Xi ա;Ԇ39WPJI66ϦuJkݼd_\qь8IC.0(:6y-]mY^czߑ¢K /-[a#:p`( ٣NC0v<L D^ӽNWtUF*KlU_a3U0H*Xsxɒ7لWO乔K|ӥH##pxfPU~K-c$P!PtJS딸)Q!+[ֶM9}B\)PUk%yHن l\t]*Q)~7~|g@{-'3yo݂4}|$iO#9 {Q;_>N^sT*pu hieVbmkWU(odR7:pDe"+9$~Kˎ"l%rCRUֹCZ0XkdMcM &NNLc_ˏګ_~V\T?x9 |.~BW|%%qѕ[r 9JLX-M!`92x,.?\pC%;:`TQ{gvopGJaM\{(\ug|Pu sځ,ȐL3ѯNYyqYI%ӫNQxO+1OP=G"ׂ>ٱr܆"œ,0v: 2fm">{>ZTsf,W D0%-p{EQHk̘@Q)4H|:!ouj|L.x;+]Hۋ‚^770xs >ul ;UlD& {::$FFX| |y#0p`HE\(xA1vƢ+V3*k嬆`KaݸeqǖXlXkPwq&>=V'Wvt7+fH"9;,i -@L:brJ bRArY&Y]AYTU6 b:YHʈ<}jz&c#8)4t&R1g9J%OȢVl*d O1Mߟ߫8b8ZGFg?eW)xW 0pw2ʼncU" Ta63WI\qwVAaޥCG.~.xra{?ad'^^-0JoKkRBI^G&T]̑aVXA2 n0~o E3-5-8zu݅K^M&./U?F> r=hE sۤyَ{޽ptCǬ`yMZjĆ6߿vM ֮gR`q1)+w9cd,q6s93{ےɁyqhߋ 8A9.q|bG!>7wذDo?c0v3͛`EΞ-zkn>L/g ]bk4OtVp3Ƴ/W6y3ْ۫GqP8>C+ NM7 7=zX!D" \(xt*U,X!)UX~MRjIRdp#6I{= G6}㗡8w7gCP}h>q<`RE0"T5Xky) !:4Xs2ˢwwUϞu\c;ޑ=mp;t=LYwvdZGbZҍdkqҍ}ƾtc_/ؗnK7ҽ&3S*lQ(8s54\~C(ZB2(Ŭx?nt.V˖5!;l'T0i@!V`WX*J5DB\Uk1&bɪ"64 vULLF%:Q?W;}Jl2-gtK:lhU Oμu=caJCP[y]1R̞H17 w>LC 'saEl8!QC Phlˠӭ O߼C}gn; gsJLɇͪ:jtfہ-wSogPdP4!aQ"zc)DTAɩO:-J \t3bgzMOgy;@>4CRHa*YoKJ< 7!j}N8SV2Qv߶&lA|j+sՉz{=, O׹X~gO2hAUSe[1^^\p[U혿Lwztp>ׯ`IGCaE(<(4a Fx0} e$c%N{Đ'9B)+ 4ҖJ#x `70RqǛ/ 襤^Nhq (,)ͼ17?~6/ϿZP~ɲ0=?bja}D6#Gfײ(+$Z_S.|ʍ2L.ߎfwMN{JuljػF&k+#j daPNјj9MpG#آ(NJz(A>w`ȚY.."?7^I2:<=ɫ,>rv~zCZz^~ؕ%WJKt||y TPh-ry0yY],P# G̳wppln.iԦg1ua#y`I]ϧhR/{_#U7X԰5*\T^p9>}㻷Ow'~8̜^E0 `1C붆Cs#6 i f\#7{$[KccZ}]$*Nfu 7*ޟBE)?K%*҄M~kܝ]][#]EwMO?gWu̖T1JGI~;e.<.P~9'Mϐ 7 x$1+#1Ko4wJϛyDg3 VTi^:(5QQpPWw$$:Rޑ}.JU>Ӊ53! =dE,JۃPZB R΂͋>Bf&6sRF?SuOr괛g>/3y:a3|*5Pvћ.qOܣpaJ@R>\fAIU0{/a8/QWcb4,EOF_9 gМ'dJ?s5⩘+Vy*KIMgn[dky**K\e)-34WBr)B}:*S1WYZ\e)s4WpvS8VKT{W=R.0ޛ_][FW/L<7?&Y4lԿ# {T uL甍E{?KŜZ{S7~m} 8*H,>85蚅[5p4C-VER*Axoo{# C4dr&#rpnkyemyp)!RbLqw#^5OxrMڨݓ)Qѓ=91xJƼj I`wjR' 7ˢ4о_-DfdJeiN|18؛W) DZ/8:7ȧRfݶW gΒA `l#crɆ D^$PVU%ZKA.i<ٵ6גf'rJdt&M)_-~ W;R/h.j{b&.5[F&q."b;'=ޒ{L(s om}ky٭ s>&r QEV:j3371BKI9 zZx[4_oJ|瓲?2Mm}z˧Gg2-|/(_K^f"ΦO6qttCm&mv*fw(a '#Us&)&}3S]*ԕڐt4H}T:D=ONF^|8k&D 79=D!$G&;S6g#څ@ڭcnS}l:bQjRvEHg TĄIl D.|ᮅ:,`DVXTX1괯 .E:BjrFנGyӹPG7q급BTmJ6>ҝ>E3?ܣysyoߟ3K8V2LIp>C}xe!7Q^zt}; ସTņ ;C{ƙ@= L/N\o"^yz%ؕ&6" X ;<9-DPU&x}{FWOv:dK ʫbyoᷣLvxeL/_/>Hk2iO|%2]/ആ(E3Qh9jkbQXPA(Z]dvE˨*XK {cK[zCQja-|G6=9ߏs>A5UQ%|(!➻lؕu'`d{bYq:Řu>J=,+36Aed ]ۼwh`I:~%:9RP+&(=)٨S>k/P#1 Vo!M  62u6nl7T>-6-|!t'ry1kDN/jXWT޾7 &"B-\Q 9vDp[Hb|m]ԜUC(j#()K%koRꍌَJ7,B:capXx?ycV֗'kyNg旿0dir2?䢷UDC\\:1ٖ,;eH<վKV'9{es64شɻ]A VTո"|%0^%vr<ګ덨=2 O1fr(.( X7R!%J ^؜B W@[PѲxXcZe /,`Maբ&(%R."qSꌇَ ~׭ b7x,"ΈGDNϭ, pϢ6)Eb.*D(bѣhLЗ!Ơ7fpV*w gr{-"[{Q+s oY>u\{AR<mg\4#.x'HX*5-ޟ8ї 51Fc,h+S`6l=\<K;3 Lך5"ycxh6ip:l6=bTNbUrvdBM?Gd6.V6p6LxzOs&Q)WLE^g9ncrʼn/Y8GpL>D]L{XGq:© ]_$g8:Ins,'>̦}^LJTkSeEka fԬ%,9c雅V9%p;o6lʓf\w|T92U/r0`+RSI9( J`gP+H1^ o>ܞ}HwHNi07G'\@wu{2$gn n$CՄye˰k^']+ng{iY}r Zڮ[6n^^LƔm|&a$}Fk±֝d iB_#|,Z=Y4cLWbAEg,l1ւFzjdƘUyb8?uMkv uXk*:䉚 $Ph[hWě֩(׹ zMs>;p(btuÁ ^pb G%x}3s=S> hϵ{waˤG*Wj0!XSJ1k?\Wl\{{c_~.K۠5(9 d˾h %hur/ō5XthUS°r@tΦe\\r"K>Q\ "><{aMծ< wʼ SSˏvqlԃƻIuӚČ|wVKZo[{[gqBZi:+ EvB,Z/rOi*D>&3{X_[&.'|I~]&%^iJgb ɼ"YQIlb&% f͒ɁEjhX0sZBWCc9/˗|Z~N͗sH6s9eyh߬(C-og%?-Wg=Qkx$cm˯:m<<=]a6\?zT|wƦauK I[8;jZgDe׽?T8dTc''ls*RV[O.PET2qp.c"Ck SأD9CRHA&O`_@DqbskM{r8\Kd*j]w$&ѧh4&VhVWJȦʁa >56*L:DZ#gR &`K&faEcaIhZ(0|JecGcQԠ#hp&φkJ; aq} !fLmۡ2~eK't2&+ϴzL{ 'mIqNZoj*rr:3QWN _?zp]ݛed CA5!6*Z]Ic`CQQU`rFRs~Υid ӹLr_VL 㽣beN,֣cI*s %l,PoZ_Ƞ-Ca9!,7'Oy}ΪncW&KBcL앲 M+J$H9"SjY ZR!];~Zl]3$9]]rwmP3D]mo#7+~]K,${ 6-6 8\0,{e{gf=lIQ[K3 n5SOɪ?jk0ÇK.+S˝xJV<-ѽ`]}mt$@[J :Gj-u: ,N]^ώďז?يU2 P*1h5!Z/.߈C!M}zYQ^5񥴿hKѴV]HZf\x^D^_|44pZM]hru]tmVy _qjRr:ӧݣmvu_׫l0[8 Nurz|2׭]Y(Mn@>=~}7ʝ$DOK.-YY_3ֽͬOL;'?N/}ys{[UVk7^e9GZYj~Q[B8ꁇZ/NGǟͷzݛTׯW:Jc8-?%h 4[667 o. vL27viяレoN/f?ӭל)uPQ^~bcͷ^T/yw+U:08ӈ@ߺmܣ2(I`tM<.գ>pzu'R:N7K2^}7}/T)6J"'M W]˓/UR%AX]4')R`!c&mHxJXCP,yMzd S"`oJitSx»"+b(V xKA^BT.lIn62*؁u雍Ww9arjg\bc|b'vXYJ TJsu\ Ʌ9!GO;HQbMpy/ѱ[e[J(xKJ2¥$Hqآ ŬGI,F/@Ԫٻ*u2Ee戃e3]*LDӫn'zg鄋4Sg^.G'[!~ۛ|ʝ-n{-l|Z[Mx2P;e >r1^?Pyc/ovA?Y-v;~bw{mX$k(.]Ui&[IJUlUq2iKl Jj1PuvbTު@Er6  ?p>8[bvs?\mfi.OyNlUnlxf5Jth]tAM"֩|N)(CڈUcg͕ h/>P4.(uXdsJduIs %g}o͵_w両MOL1%CdBjmr9(4Ot,(ؙK;.Ww٫!RnhiC1Z1>:n%BICP&:n@҅m,k [O"1x{v_ˎ9JKLbf/|2Ȫcq[g+8(Uٛ(].*ҿǞ뻳LjnoM~p>){$whr|@=)k.-&vg|k\XvRd`*yKnkǭnq,^ii+hLfcñ7ؘ:r նߑ0X8A )NKu`[a[+ە}=fRPC BHLI3բy!rΰ 娒t҆aW>yaq(wviKsy_}zgջ\6OfY@mdFwV6:;z!vڠGcGp瓲$Gܘ4 4jA WD?ʨHEW>5 k>Db%ir0Z(xYtrU:3ytF{NdwXzJԇ ,BS~t;oGm]߹+_ 2-<ճV$X 7ero7iMJMğLYb#a(Yljգν \=FJ/4\G=S^:#j[7pu/pդeujRz{#UwCUUΒ{W7pĕv_IU>C"Nj{W*jjjpդW#\y)ܭ\ utw”Nj}FWބvoAo WLye+G!oϯ2_oK53Q<-)oAa tžtf9´e~`$޼bhjpդ\ytE6C(sjvzK0Maڽ|-";Q+?h ~6iulRZ:g6,"6qiov^^~p'Fj' Ώ6+m,οet2i=msts[JaU=p7n.La|JEۤ8wqd|yu>[=mKbhamE;=EHW{ZW[̭53aF,pcwt!Kemg~!vx.o{4( uƫ>@fYE@r+Ê뽟y@x}~G {T*F㡋~:qVkuFf-Ჶ2YS-i pKH^{E>W H>I;/t?{4+$3_[Mr:R\ V ^V=Anr"`ʮ6$Wd6J]'9Ir"$T(CZLV( 1fV-|ɒن_a$)[I#ze?4mʙZ- c9#ۚBQ( @Pꕖ{,Z˜ɓz[J}TBȘC-`T6#4) kR!TP2ƊmAgc*h*C]M@m+YDlfBz-@! A,~+W. vTh&+GXyJ'[4~:VPfͪ$(U9P3FD+IAfv+ʢZ ^ƠL3Dؤu}!p: -r )1RP+59KDs2Iecv ًLb>I+~1..L!)Q-4[YC.$43ljUBV@jV Y\wP$BrwQVm)$2*KV^ ˆ' ’-OYEF' 9+0OCۊ*եqQvRX˃mnDc A_4TL@sicnUMOXLĎqAP iXgGf3昘-e]8A`DNVQ%KV4ZxÜkoʵ% $Nl\9 E^!(ԠRm5 )l[M&k UR%B1%HF3 l„yA R!ҎVjd=,g0!.QfoU{C% a`\ ̽0 RX "Y2Mly-}%i+UKdIv"SUqk@A{ŜmŠfP*KQr$*R!abD _XGY0a=XgҾJ "l VJ!()sB uF#%\Ӽ! B0Gjx{X)B6 Nr@N)" uJ.9eU߃.P˒'[#<̋U BB)X%\LpBi5^V+a$N +8oVǂkFM'23:7AN7ZKo+f<kcD%ʼnԨ ϛ}H0,lg;nir}yUfe߶1.cVP 3]Swu yYT@-$ >^x:pR?GƃV%C(] 6Wr*=h`j6xLF'~aCE :.%%* zHXDunP& :U(awFR8-&sJ$N <&ڽy/69~0~2$א*b%V n,zإt!zŠqĆpT%ow!|9b(#Tu&mQ< ,C)906y"Zю:!w0 Fd!5 *$#nH{3(g4`b- F2~/ܰp2B"mQٻF$Wu>E{v``k_f0""`,RuJ hbUTf2#s4pqإl,f$@frP%JI>2 )eC "RFdsp0",0 E*`'R% fcI;0] -,f5*)8Y xJUΪrRo^" ߠn%LzHD_m,mPҹdݦ6i̭WWv)NQW,4l4*-骅 $8z`2o6:7 ۶{GSpRԘfxk]Ь!k sQS@p8,8Z;f c\`ʛq@7U𙻶 <%fM:3fsC < %m~D[9,B2a:G D wB)W ;`X746I2'Ջ [1L"\x>W' W@Na|;)E^g<[0baBl !ZXcx)U,J#jOZ7zmxAנh\ aQЏ'#b pNiu)Fndj3SkHkYr`v}ȤOd:USEc?e onY#9aנ\U8/ڒ?샩%t/5s%Z6 -XBa,gfC8 XV2#С /j@ X?nL78{ =F!le(14qFO* 󋆗 󛍵9mِCLXKo&b] cd0+&- !"6CB,9)`$IEkTˮHx;@3- T GeW+yR絸q:hˊsڮv1ժ"k%9K}EpzmBZ#:kespNS6E/Z8ngGݸs'v7wܹ B}Rw.6sgfϚv?{J 0H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@zJ '8ۧ# ڽ9¬Yw`qX)$X@I)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@/V BOTyȷ %P3}Q5k%u%PR RD%f"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RU f@E(ښy%@ KJRJJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%Q]-G}ciz7i^\,;g4= A k> _p 7%X+ `f$\nK]gc v#\\dyNV*v \ݷ9ȕ{W0q7pv_ zvB\@Z)Op W\#͵Wʵ6^\Im{W0(7pjڝgWJ ^ \j[Op W{W\!j׮ W^5b}cY2tҕgϾ9x퓶7͖Y u-0yo 'Ҍs;o8{#k3Vj{{wS|4_~0 _wgZL=(=i@:gy8TD"wj%Dŷ=K3أH+?U䚵rV4EܓDrXJ{W`'^ \Z!URjWJ`-U3W \5kγ+XiՋ+˥8_O߷nMF%YsK|W!>8_E]E6o'mVR)j CZVDV3h :&WKd>W_NߦU#,hHyeb_^3+۵~[`۳iOa[!E˵"YBSkc~4TVV3_@z"~įtKRﺈyCv-0Q>?Ǿ p`/co_֮zvН'a|<|y_t|\>ܢ61;<88N81}[iuwDbޕ$uY (]qcum}LoKe# ^nduf"YBqTJ(U;[Y=~nqtF9x:OO'>HPZIcuE)Bg8WDyUԅYce!#G>C#hA[&9%Aܪ*MYRY=YgkVi)nmZƊ44 ̜I0d|TOkxeҙl UzSpeMd̺d^L6ⳎJ2TmІ5KYZZ]Vc*KTB.ߦOW^ܐ&0x{ +4{kH3%0W9jY+ԮҬ\8v y98/ɬ?dDn6ӻɷyj(9&u8][OBf]YtdjUbgӪ2mjþ9[캅e()|6*#=TD+֪|)+/e`Q!fk٪\6$). .mKmQͺ6>ygvؒ2˚5755G-Pw7gHRyҖ/k 7pEd||PD<gL<<:';p祄`ֽOU|+IN)e3<&#cN۬S]cB뼗-HYG'w&e%,@j. w6i&W1{)M :m)Vᘋ樸9ݪEe9{[ 5/hs0snf=1 ;DC%߾̲8o 4"p &]ia,*)\5%xY81BiM`^0o"7էUx**bsԢZЛ e?@KhNZ:F5u ' 2vFVrErt+y6oy'󓯽Rz(izמ's\q_̵7q~/E͵ +K-->uZU_x^Sa P%8(2gk5FvL1*$(r(ժ,2VF̹*a`qO,l`,􄅏+ڻO3P/dmg_giA x4]-Iy6/6Ei2Ū&ejY9W[bx)gպ\pKͼ,BUR"d0CE'Y,&VKBUgT{y`܌}PVۀqWq0/"""nr`8%dLvN1732D8cƶ)V'l gbɩbV"i 6H3]n9vis0//.ʁqQ..ne!dkͭ ׎#C0M:-^y2ljIQI; l%<^y9W" ˥n{㢌vEU?>Sc+~*t~4'ZhmRGƳ!gY۸+S]F*וxשMvQK\*<%)&)9ڭטI="4Ʈ  >g,\=A;>]Q6sCS`S0>Q)gp&)Vk`wkľeG Xx,U*Ԇ-I5Ae8)am-pU.95>fq>e\כ70)EMf9%liCz՛*fbzyU˫q^'o)H>܆*5b{#iΜ*sq)Jt'c\ q63w~ң`ұ(!4i魰2TBZtP:Vtˆ^VxJ# be}0z hA #(HxlLDe!5p JYlUq)}{MݜAJ],O?  zeҍ_\wvݗwKv~ i1و96HMk]4LO.Wh31cpP )`A 8(,<5GŰ 1 h13 +ݬ97zs9|!ZSI$ywv?&i¦w@Td ּyR-1ZIņ;#!X< FRt(* 3-|zYOy:uW 9HH=9"Q# ?hGcdC YE Wyn(ң/&Pw1 sz HR,r@`Q*<H {po48=շF+zO z]hm6$ NBQۚ tL-Km0DUC@C>O/]mјBEHvXE8Enj=^F$&8ST7ʹ?aGkr-o!ߞ,*Pq,m2Bq.Trl(Sb1[l÷.*1 U <O營TAOӴz0t:pTd09>Gh@J@m}Ӽ&տ"|~1.ϐKW::oz?:tz[iVg^񿯗% t4Ŵ+sZÍNP)Ю\y^ǹnn(J)Zvi2B!Lj~"yYդS~DI59VָL1}s#λP@0 &]g}&{\; m5?H@ǔɵÄ(2ȼ6X*Q\)@@,ZCJZJ$I H̏ Oc*5(K)µKzq3\Ζ~K<}b_$Ԯ=^$H.|1sn?cE>,!69#Xh.]oH;y(GPDR{c;AS4}@$p>bJ`if*')I*WtL*"QMM+!H&W;ܮX TJ "`]tJb/M۠0.w秉g&/n)\eSS0j`gԿUhP(nj$2Dn^0e=]KTM\,|,Ds][S;+'3#Z~-[> o/gօ;\1[W{f(T|ݙ/7woB 38wt4 iFi֑чQiM>ꍞ'jͳ֋li&Gzyzz5~>\b+f">8oXT  es3@?w?|x~=׏}8yuw'­'؁Epp16 ¯@Gpk꿼ye[S|Tj÷y Ɨ -ւ(|۱a8wK~-^ң,Sd>WoЖӋ(M+ܹ/Xa@k~{F4[|.iJ7e$N3&fX< (noI7$(U`#)E9Fc~Lu+u61YQpPKhb 9Q1MEE=uyA=^by)4`CH=EI5s%znƇdaH 8ng\P5={K*pkzrD×M-~|OU_G(kBsu%c|qg Wkip/_f-\Fuc3XnX"(#-:{Lo 9 sIH3S&M@UE8=s\bGe?XD{_N*'5 j_Ӡv];t]w>7(6 ڗlת%/ Po9ǖVdaHXsVXl<F_ŘܴJ1ZleF$ F刊`) NPD2-SpkOwY_'=69,^xONT%^|!R!DRw'i3OJ30P/ n2Lʕb 7eAh̏/"lEJ3qkmD闔b.Fx=-u A@dQȢn "R~BY:8*U* >(RI/` mwQh.ܷtBܖj`h[p\oj]GYa++?u.NzlZ4wdG0Ao?f%(äl/7gW& 1\̖\}9L>?WI \ʫ &;(gDXJs^Ju&T+Ǯf.=3&J,$-a /^ ZÅ<' A햝u(joFI pxZ82(B)&8 n[m :앧#@Q`De*ϽS 1^JHH 9z/RIAzJ,>bG-$b`[ N)L W˜%:@@BCBO8. åh\ *"Sb#dG(jiRPX[zN(LW}rnXЄlXsLrdh5"EΜSAr"@K< Vx iC46 ePyG%^ŖjwH[PAEjKs: ;(y׋ee_/vy JzvxPC+#vjoU}aW Zή+ThHW \]%h:JP ҳWȮ&=bW `?*񾰫22DZ3Q.>f+TΫGٻ5[]a[fi-eqzhdwl(zNc#!tD҇i x ,u3L{æR Nγi^ ~lk>) `?ReW.Ae@I+dWBk-+ ${îL JJuv=zJ*ʁc4)"Ml RٛU.TTd)nr9^Qv;|y~"5b{#iΜro,sY9Cx(շ$JO~|5vrF/?˷G%6E]_Bk:0m [I@1CfaExKe 5h,S/S0bH(,V #/M),iv^ɵlx MɠӅ%;(g{׬m$|Pc C/SB<$9Ÿr/Ar9Aw $8tWUKiQ&%{Tϐ2%L5%9atOUWW.DPH(`5VUil"9469*HdPJ{axUP,moi[m(V4) J2=B\V3qv3BwUS)*{¯~=OPպOv@>$܃P/(S~DV!^zq!*ecH; :tri`x?4Nr(1EȪ#3&Z:!F',Q̱9,d/٣C22de <*J!['+`G1Q+ 7Cc{@yҳ/ t/}L`ک$W hl V$+"stVِ*eJ%RR Lf#J[.[ދ 8 NًfwNcLO|<1kU֑7A@mU궝>XSdL}sOZYO\c;`` 0Ln.Xr&^-CA`=Z,9tҚTZ#c3qPWi4cO,T1 :^]Wf|Y%.>-`هeti:}/q.^0`"~-c2YОݩjg2j۝٠PJ蜕V6Lv v"ĢL$>ޙm%nγ \s1iǑ~#j v[3WТ e76%=` #TTRㆤlS<,ooCff( {Ѡ3 WG( e,SM>2Lx; b3ecD#"x)3áCrX'!$+|қQYeD!ڱ@m[yþ1Ye,MQL #'2^S6FfFďgSҸ)LKEՈ#.˂5;id0RDUKЀK%+Hv0!phxb+Mc<ܳa &+lW_#wF_6KnڧQ }Tw'ϔ[A*q5)xN/d-uctWii 3E3Z`5Lc2dg9S_b%S!H d QD2ɢ31C LjNJpmIV5no~I(;qFyztV'LM/&W<.zz2{߽oan"Wez2_pfp1Ŭ'4n{#He\(8mnXڍBqap p!KHRYaU<Z:9Ƽz杙^س&%l2]BНI"w9eՉg5;g$P!Pw,d/zIG?Y#$" ]Z:2ȤeaEqXb"PGڨ`r!uBTVH>x`ͼ^[>{q9Zr3δ  m^GLԇr>&d݇sDj UO-$1$< 2OM̓ }dHxŋ /cNZ6A ~MOAj?4yw&勶 қ]=DmЫ}7V+gJEd鳴*IP3qv j*}>l^B[pDTF`x`bV{`- %)H:,H1r1d)&IcD٣iMigK4- fpTk )#2E"-iB/xR 2Yplp3MjbVv3sxl~u#ly9ʣRCXkp>!2Y t%"+̄rōmӪwn<b<ZgFolE+deswl)IgEM DIA9d1\Yh[6b-C%C*Rڳ;ϻR -B[a GzߧK*%D8$$䔃x(6)`QX)R^*͢\E>߾SAJټOVku}8ǚ6SRNl!|}ߧׄ9=:ίa3G@>i~_1/vlSA6)y\wGfCC)2\re]_)q |t­C=-]ѻ.;y{tKS.XeU]-/h9jz(ws33+/5 CNӧ{h/WNpŚ\h!# F@֏ NN_x>M?F @;QO(MVQSp>Dw^TB*S1:,pd4 H/,F}{7ls=`$Tr5g 2E]:[)D.VLh"^f/U r-Xj7Ӎ}#ږ=VBwzg:qӡOWO'@~8lrN)2]u|Iـ}@u_o[}ֳJZ2LZJ0c!bwW?!?I-H n6(VO3szF?q:yXiWEK'_DWldNhOʼ0-U;W[gf o$?苧u1fJ=Mz8n;{ /v:>udž%}~7eZN?p7q V:z\]//O ;kThR&<_nvO^OX'fQkMdIWz);I;'[>~-  PtFإuEy{)Y|8T}ROsIe{9l{f%BVJYBvc l&7g EL=£YfJrȓ04~ڒ𶺼-m[NJ)[O Ej ljD*>!X(F@lRolz/Tlx5s,@ ) ґ،>ʂQ`y`/)h)c)I:"%gfK*8] X)$+ 5C |d|MCnJcH$CJ)=eu?HdBTUT)GӀKʽWnwnc"QZT>cWHQYu[E4 e!$3؜+m;ަqBnSlc[u`и{Lu[&X^* xF?T=:E͡ JDPv ۡ~W^.R+m EbI@Pl [LQE #r ;7z>ߧ;ly*չ,hTA[-kgF# Jfcaؾg@Zg7p(`<{|V.5t1&Tk9mxT Z{@s: 6Yum$Kp%;!t)k&NJ69c"*,8=0lJȘ0;.y JdZ՞ %YP0R[Ca9uPYCgS ".RfA>$_OqY{˫ٛ5"rJW>u +XNX͛YL~2䲺Ǟ ~QIQ<-kknHĵ5dͩ}G5WdHɲ$E$DiѩXF}lR}*cH)D0u.S1q (N|r gE-6(k kj9Ԕ\6.TEп)jbai; >3Y??lh E2%o Mtv6|cFHRכktLYliJӥŸ`+ Jޚ&wg)W~..|p>Ζŵ^[g{;(Vzy !(Z;<A?(t`djdz9ݛHJYEZVFSjqz#T2R`зc_1KS٠|aK*bP<//g;v//ޝ]o޽D]x~7`u  C5 ؊M ?e[S|TSS\dÛ\rͼKć鼅CuSZ_oql>P2 ^F=iRoO^Gf|`0(-ƗbMXQppKhbQ[ϨFF9FQqt>:Rxd2KI5kBCYs )CqcJ ;>j!^bj'64Mdn<62f&^ǀ|ɷR zm.?+~뷐a%Q,ؘTB2:V9u,jFE'xtDbzLRk(RmkADY%3IrypȁvۄXLHRVa S ޙ3@-"﵌FMFS2VHKD5r>nާyߌャo_7i(KgS)|z/{11Sݤ]3 7z0)Ɯ!Xk{,\.WNi ńV:X4kP:(aWmg7˥y0 VW]]b x~Џ͂'W M.3&m vfn+S֬_4 Z_,i#KRKQM'!y7&<%P}sOt4D}ww|w.:رϙe,A$8G}gs]w{ǔ: =e75e\US!1H%v*pJF,B0`<)V4hA #(H8H컷FNC}dˇ߂,6}bs]Ca󹹽*in}d/>UX z9g)Nz<腪lAQ^&0:u~m810 f1grH8ʬ GstC]ztEZڤZn z%mWc4A0ɵ r B($<\2# (4j$]|x>;{eeeS}\?\Y66 -oC$+೻nd4&P$ua ]4uG_]&-Ѿ,NH{[9\(.6wv$s})$HTCf߻N)f,Y-XKc zg wt K(.MkFu5jE*,rJD\Dc)xw:xlJa^iGjd^ 1[,} *{z^QCg7;5ghc{DZτB iM4?ȹ+_6Ж4<~޴`6(8Ƀ6hkʹ.6hLW xdQ D`!7@Bj|+lBY´@sg0i#yA!ӳLFyɶUjG(gSMCVX2>VkѽktCt; ij%~=s[Ǒ;>-Uy?J-~D[b}lKcH]%?uȝ@sutURzJKf*c; *fR|\~ 8 ~wUZK ~vim.hyXkCXFRcm;0YV3o|^$jh,T7d&>zr%5(XIڂ=k|n837xpZ@!OoGŮb wi9G%s2J:I:F0ΆyMb$9b+/Gyv/t`Z"FIԏÛ N{_yJslۘkv<;aF,X1l_eg§Fӝzs3UA<\QQk:[YΤs^|A/'/oJ5 KԖe%*u|Xhbx5#X-bCb#[`u*Ij/9v3|%gS>QO'w9L0oF^[M,sj&?8tG_iNQfC6xcL i,gM9 5JL%pnsJ]Id5U>j(bkf݃t٤M,K)ĸw7gQ^6hl l~'kxdB*deezV^*_0[-qMLѥgŒ^f UήN(?zu#pjNkp'|f|>U>n4R5}9-FfiiЬXm qͰkxٜ%7e4(Y{;\$Pd"P_vEYhKCn-)zT32M1)gadr6%˩(, 1D[sW6tT UV|FED#vhvӐ3 J- N­!*) `b4&tpQ/5VEnmЊ{=PL$̎Skf*6$Qfo{pYO䡪KEߏ_%yѸ?~1k q6)&*Vÿ~8l!>Kb@}D^^~\VF"1)*/k>Dۖǎ#`@<Q:7G"RrKߺ<>|MӷA݀}ZhwbEK}׃<& p`1H9c yapD.L1P@8q čE ! ( 0Ԗim-ݦ|8uy;/]YN%]+pe^A]#wG{E."ea .M4|(0U {O%[?ߧ+쓚c7(UKAĸrD $PwiR{c;ű!@cۭיB]uz7ZJNc<3̰BWN+ĀII$[|&=V8(Y=E*ٞ:⹅Q& hB4F'SF L Rncc'>#? à;Ay  @ $!P Yh6lJԮ X=뛊ߟ bya[Z/F z!)Ϲ p/ҙmP xsC{?ٻ޶rcW plsXۻ(~ieh%GR8^,:lӶ]7ѡ9pѿg]MmVpQ<ԇ[dJź#ú@qN1L_Sk2/x2%#xq̟-WY9=% X1򃥸&<~FX|X?F>a(6^]֐gf/d+Qd:w!oz5)-BW/WxY@YpNGHmu͞>lUpAݍZ\o<8?=ͭt7s("U3?\NQudf7.ÀfUYBL86j4K]vpp͘-/gt*Q/5j׻*)[:e5R֤>ې/,u&GXĖN Ŀfg~?/??#htݼSx=J'!ݯ[ -AZ2Vϸ+waʿί,vI/~OfgFC]s2}WwJR)go @a@/F\s NriMlIݶD!J7o1|Ɣncl:cV@ vdb)2S=DBf3V4hIA{9Jh`<$%,!6+ MO;MԿ?'+{r>-n?ly}t~g7{;:NhwJh~͈svfG [}\~%.@^o\e_4fsԦYo]{`$49Yl4/dc\l* j͗L}s;sP̭w%+QsZ7v- On}>pׯ ]lHF e"J%qww {.q٣MZf%OjI"QUFu\;Bjˣ@dCS %L(P:+̒CHL%i0 gW=H䐘 Uٓx-wBs\ϯB}ej4C0>ڬD p5lb&LAk4뙷NGeXK1OLjFMRڑ@t<($|!CskHO*Gkv%gu\{'\>{f9n3*+zfB̲;\Y޷I>H{RT=l,;FEZC$ox%k?RK#S!c<ˀV S % U`ӁۄN˥ pf !Ajm|Ԯl_iGHF6Mrkwoo]_puv[dtQg|Ew{G-\JσAe(yf[ڰѶP{ !-|:tk>.6ZNs eF2a~noiG[A[++@ro?͞ Lgʳ> qFQFu%LV "^'9/,n/mtH.Y#EҖEDžF9H:8VGEv"5ͱGG/8ƞ 61 pڵ1h@4QiלOFCFC7 *V䚠k"D%>m*֫wGyh0J`Rq%t\,(Ǹ9vPO]{Nn(ֽ'EiohCY{~(ί"=vsN:GDFZeRQ5堆?%0v(%Yjժds Wa/ҟ/Q~ʼnH_@c+!@*9W/E&мQ>Qj3,4ѰL?,((\(2#c2z[A@A LlRv-UĨ0$z(& FL,kAˊ2e )A &GW&WfWMy1=ÔFW+,<뙻4x6w|qD]AIf⤪'Y `U,FkZ23B/G/9z{#iO6jҽ1x.Vbלmo^8e/jq"_uгȓnEg,?NyG8I4G)G6(M?g[SQpQӥ$g?LRd*#Tb+M6ILi1[+rC9Jhʑn,GWɶ4e}vۧzDȤ3f||҄՞PO3m4XT@w1R$6hC@́,ԡNu>H'8 `2A:6 OK:FF$#DHDF`}:Lй@od^~|Llρ5r§8XPR@9FxIqBX7 b LV@vu0k^Ћ#@ %1SA8*۸=o:zAQiR z5_UVʻcAJ8VZ8.ڂV;V/(\X0 ir #"^р)XɘbR.yPaڦH%nFôB1jWӎz#΅LN:)4-}"\{σɃ!4Ǭ֙'49BWYc1dB\-ٚxDCl}C:H]ƨ8aK0EEE-V| Oo%u^r\H\L@( wXnNgXEeI&D.v&AAíDPVRE$Hl3:],ʸ<.vQv[ɚlS oˤk :Π` !5gL]| x08v<P=> [^O*=ZI6,e٨e HM6?h{6έX?x4\wR pFx%ЃD\ePM=Lx֐G@.u:Ȼ[œ_.n_E]AdxVm^Iשi3~îUR &?= T5s~#y 6Ͽ˝䎹{bcMg?lf׹QwύŻ0c&efji2p3ZG)B7  X+i}Bm MĎ3QR"Ķr]tՍ*V܉{Y"B 7~r>%m>4X҇ZP'h!jT$ɲBESΦ^-:UOwԋϝ75"E/P$~[ɝO= *ǝ.}+ו j`*߷|a%x{Vy=,\`gESS¬{~&g`s+XS58@JCH3B%DӠK5,kfvrÓ$ѻAm05»,_EWcwfgnn/ذpY묘տ~SoMg o:Z-PtT5{(9YԢ"/l=˃LN㏀bNWvz0#+"q4pX*Ip 0)?"GWI\JL:\%)e3|DpVh*]9iE$%U=\BbLKqL Vx*ٱUC$eW4k[|BӕՓ* x:Z^VRV ~ſ~WLdzL~K (;C4E\zvsik\emD1^: +v R0:FG)IAM-uXęP9#\\}IKV<NR*O.pr)NW;afF0ۑxR|'Er8t3} 7ܜ`k~) (i/`̕.>ݾrXO٢2#oǟ~>c=݌)rTGa۬a6 L_z_5"T fj& ~Uv6dp^{Ǜu^IM[~چvj'4@ߐs'Ҝ ]qJbr1P,VΦa ʱ$QK-9+R 3}kwkWG'(kt)|%;Fw]%4k9|`P^iGjـjuwE%Kv>6`j(*|?w1 f0"K3'8>ԗNb5o{.v }36>ƟFg FpɀŴr^,J>7kEa;UKԥBMp}`m|DJ{5ڗ(\37JO`u5.'lUi/aZ_d P)+AOnY=ihĂSIi!XQ>jdS,eD9,Fҏ5sʍE"VFudXD#aREbz6q6B@vɧIk>qsm6δ -& =< eިʠ |Z~~HPQX U`hi)1H[$)$= bMس IeD$ LZ$x<#5(LAu)p LUD~o2Z&-. awPWMd* ƨ6AFe 04^s䬒l58GF!7vgn0oM},]bj0ؼYa+~}BL&~aWWQX8"⏟AGG?궿]F~6Ky/]ji,mé?h7.oB2pV4:d $OKfE,-D'M_v[ԒCi_1}hMsiDe&OԎo\XΆưd1W;b۱_)Qw">Nї(u XS0U*Dy0e:))vD GOv߃ (-N=;ɒ텒LP-itV9̸!t 3MDP8rDQ|E \ACNyBt,&pj{}k_ו3(g ʸJ>=ڃDJb^ Oi|z̰\ys{=%9Gm 2]~. %=ԇ: ;`E*°1aNz),wqD0/KהDe,.C0l$ )0^o;c-nY$"﵌5MVЈ̬38)B!%/+/^G +/#X3eYF<:0('Jhg-8j! b <*0r!,$S`fR"2j6q6j3GSLP^eӓ]^Ix5MoaCJreޥ}pΥ,b 1\C K~\ `0h-R$|H%bd$XP1g&HFllF|\%fӌ<(  KRf7w5/E j͆Uw7J3? /?O3#v8:U,JlQ"mD@?#cbG-Q )ejޣE#nH= JlR!`V0/2QKǔ .D}ԑK͈m#Zb jgӎA]StZ:&)=gVo9̍V:o6$ri <+F.qgBf vG(x`# Ɍٌ~ᢖP1~슈83"{D\r WFXcj bɴDVi!H!) `m8XGpz ",iPJ0I0NsōCmrn\יMKvEI=.ex)ɽ- ,G AQB9vDE N@}\ڱ+xv=@2sGj9rnQZظQގ[8ޏ/`/{?vM鎐[k!IѾ8_lKϵEMqd:|[,lxS%72c^.&6aRZK# be}0RH0<)c"ҏHy7 -,tb+a b }plK֥Azwˑk{;Ns|>]s ^rG*}ۆ%~/りcJ3&^cIw^çKRJԧ{Xک}}u]eؕ16dFJKE79w^ܝ{0{GGtT}:rYֱ(!4i魰2s Xu*Jkkg+**^XGՓ!]@ѣzaK~cɀ(fss~rG|/(OYdF̱+ː$$JS ]4LG B I@ Aa 8*%0bh^yޤ#m#ަ_0ywTTõ%W3?^ޖa3Jrs(6 aB\J0oa rm$#z q{ <3.|=#RBq8x4Y\"jdO#1!`BV($iy'uI凉@ Fz0HR,s``QJr0CZbQBYzoƺJ-:lv[8$2IG+KY@(BfK9@<UaGqo \b&FI.%*De$abPE+T` 0#0%J1хG*i.r+7V+0DDbBAT__'fɤvG]Aq4o Bq.TrZi4aH }jGl Oe Oߠ"NWޓqdW2``,YdqcuJSòf}_U7OIjJEݪzUzG1^:Ti u(QSKaa~ pt>WaxF)=r \U6o ؘO@wrUVV!uHnH`FkVMVf: YTGy1X8 p5rU,yrb;np5g+aP%¥?L>Dm^Q׈i=x}Z,;`=ඇU^8ZM?בwFh ywOmonh66uhMU6fX Iv ˖ɖa"pӻ#ñв{TEր5z4OE?2f MbGI\;Mcc+qH&7ͻ,o3@pA'LT 4\,p a5$ոZl5n5.oy#9]8Eˌ }GdI ?\PS~! ,۳ϳԙE%Yў䔇:c^~"z_C?ժ~gjμڧPqt?07[Owp ՕE6j3zo4 o=bbW MTU0H{vDSl=aӮ|chE}tAOJG!GT4G*%%^ F IJ[bCdTkѕʍB?RC 1\/v@VMw1mMSQOHX'%h 6i`z-w|L/a('ׇR5ص:ϫ*kK +gտzmTPh&(* d,R(Jd4Jp;Xb.:VxW3ztb\vF;x cBL9 ) <[~S\>Z~j#}IҪ+ LVZLXՕ0iӆeUvj,]^QɈҥ>w)yCV>z%͞[ f9ϬItGі0O{/GDD堉AEE !I%iㅑJmc}J!Sx 'e5p6ju&= ( z8`2t!dzs_)+㮇5 ^Tr14Y8JHH 9ݮ@KMR&JFD)1(A+^jFte=*ȑRR N}|ĎZI3<Rǘ1D0`TM|C!D _x0bH :q`UD4#ELqdJXF:,(w i1D*'@&P# Q[s*HNp3Y+eؑƆ@CnZ{~kҶp3|wIw`Jyy.{@Nє ^(bc9F"ҹF )(uhl3ꄞ.w?@0 ` wr\ \ u96q(ǹ, g SMԝ~ŵɌ%WeNtv6DRǨ l,_KTP^0|sU*c5_ߪ7oo'E`4s2^\V{K|oEj'6B 38t4 iFiAQU)tG9WIg%hE6ڴVFS鬣4004d+#~;TwXT5 ?uחwo^Mݷo߼;D߂ 88 FԣAQhL~ԲaTSS\u\rüyAu3ٚ ϓQu7;Mͷ?LjVa!]/G`gͻ;*\=ĝd**w3 QCVv Ƃ]{+yc%KiUgKT9(N41T}~7]52Tֽ z#(۞|Ow:hvӁ(8%VZnX1ЊQ[ϨAFF9FQq4;2e&?$5! ,J ] \81%Nz/Z:&NMf-7=01WQ_G|ѳeoLVDW&kJq}KA?{sz_T 7AR'AQOF.[uWd8 tYQb>uʲ>% kpdVrKNNUb47S\._+R\3d?~@f4Ms=jz]pfvlE`{:aze;_A1)ȶ{RDΎjvuGF' Gy\ pbHCQ':?"("'R\Mz~ łJ,P*c1NoS)Sm,'a22jɥֆZRq "UwƒͰ&1w"nE_ I$ )Id;cƌL"^ˈiDk45[!--;[gK7R]KBd rUӎo73L2`nlSuɭf6N1SZn=1=nn>+@j+Mpny9ӴOnvMwN{836MG^r=> 5MnU8vv307J ªD PsF:4߶UOGBQTCz[܎ntFSO˾%)4oߙb%TECHQDD b FрG!eLDegKkllZJo j@[gۙв|i&ꛭ:oӪOçr-2Z%4Hԑ($F{笰xD9csnjL(Xhm  .bl6*GT KIw iju[l{%6J}M lo+grpЮoz(k.]Y#vgI *)dKm|,S7W#ɣQyQ\gnMmvEORo.Wp؆1T)b+t@Hmߜ z-sn_ [un =-(lsiZZ/4fW|:[0P~ n0§lm28ɿ: GO^OޚѻI£A>``{/+MqR?Z Txr!7.2dQ,bte83տqrVW'xxOzBOuo+^ml[6n.}AZdPsΊ_G_eVTo~M.Ӄ-kdU &(^cQœ& v`fUY&-2V0Nr4ޗKW_oۢ?zϤiE$w5e 0^fIR._ ħM|@9 #eFJKE7`gjA.܁v_d0_R$9b.)2KHJ~ Jŏ/B)y5&n̒3a`بJ@e86!~x\?W֤5-R0bSL*P_ݿߤ?~$>CEιOg0giA'EK>M;O)r)E<ȝ/awP"g4u)wmmdиj6ڇI[y؝UVL4I{N忧qx$Qo"$R,[<@h|ht -<$) rAGFd }N٠ OvՂط$SΉ)ʿ:[{W[E:L^$0YG@q]"gF]}-Naċգܼ0\=N\p8iQ \=NJu`p+pЃp \q9 \i%;t*R"Tzpŵ GW$0WhઈkI+8t*R.ĞpvJhKDpE PGWE\WEZ WEʮ$\Iš#+X*~4pU5XE8x*R Wo,.fR\pZK~ܽѩr8d_o0V(vS\1hQ%z'Z̒LL$^3 ֻhLD}$d:C+RZ`BNE0[B-Qp̈x@b<; .O{g#ad%ƫ/9=hJ'[]݀'ty_|) =o̥ |I|qgJ6 TU;XFs8(D"+5{vx tirO+6k+<{ǼYo}BM@ܻ}Όۧ{bVҐ#xuclF" ՊYb\\xoh͖ţݱӇ|zK2KdrV K`DsҦ$AZ4\hFT$؊I# , tAX"ˈMT&+SJ ` 'Ck˵L9bol戄>$hvggt;)7`yϧw4~.?P:j֝ß\l_Y{k*^ ߾V5zN5J1_bZ.1E&QO"<\$EZ~H7y"wi\+6԰νnڜOރlu6xޤҚyxҾ^.|*#HI'8"z%T^tde' F>Z!9֫SrL 3YK>)b: sݕxYFl$&f#BY͝շߧQ{qu1q BqL9CkbKMRȍ4B9a!{AZ2m9ʃmH br<@H$\0F̄/H%扚'XWnR_+ >we˼]7.֥uj[tW6 r()zoVoԕN:9kp~ݳ %c4VǸ?}C; + V'C9CeK%H^R{HL< I+2*s{Vn+cb.1,L:fyt)#(PL 72&fd쏫􆅽iƎXYXh+>)HT6\ GNNCcy_hp8mx1rrV>Z/1gFb pśϋdNUeY%uC1\Lx/)Jz%6L36si C,hl_lFl? &桠voq ݕWQ2A1Jc0"M}z ,:^"64b27cxQi</B&f &]ϔ#a}|Y :ٌQ[P7gDZnyb8)Oz,\ 2!fJҜ[2Dk1% &>TRE2!%ͥN %xXψ؛8[h=&!u%yŊksQcͤ Yd Z=fl9A)x*T\| \<ؗv슇g<@؍W_tٮw=zE뜽\u~gm˔S,욭e_Gm2TGȧqp؎#N=D;ҿ#ze\r/"!T{J%fcr( E'yrF$0sٞuol9.4(`~^grX`_w981OBj$-A6xG@R,}y,6Vl1a\:~/IFD!5z 1==1 ,RT) /Ƴl _6+K@t2XpcD#w7,OBYR >J=WsAYf62K"<R%/iP)y,Xҿ-C6y! l9SKL(tH!n;c+͔4s<성&oO·_(ER>-h~4h$3D_O;l1toG~!6AH2k8s>'HWDsb~D`(SpP&S<ˣ])Վ7v¯?g|~Hy>8EIj)}/շ KKVDxD(ܩq / !Jsu -HK-1ȤaFIDS:K֋/ٕ"rp?e˔|yG祸奵Ao/[!=m Y1 Lv  )D>Eh! G&BJ(Oh SQ"SH4yB*[!H 0#&ȪD}`XkK8}a8YO= gV9I4uRb@!+-&Ksg8> u6o7_1YRsv[q(dhP"E}[v~bG_^zX=h1^;G/tB * 4aWZxF\0)#e$i#w]!%BVfA"+#HC5pAJ#ZңWֿu.FZtZDDզViIT^.*MQ7p#uuF"|Mr08|4,/u:x~%=UM.dz~-Li'zZJ|pWq82tI0'myf_U$]+4I`4p\\Ufܷc |<,bd0ס,琼 gFpz |@4.ȟ~SR굲qjخ=vاѴ2vE'(X/];%rc:mƚ Mtz0v8x픽ewl]2'eC1'9 r0T1|WRfLnWYU6E:y%ne͸ثܦ P?_LonEvTu oWڂ ] oۯrڪUoef/qqEnsh;ˆtPYn6p1ywSO++ՏU燅+L5@_|~z6sv;rC]57`4=e_j4vWJ?M¦!.r5u{^6N]M7yk'_:|{E"_]A[{Cߛ蝸J,p5/{CuME{]m֝~vl}lQ-]P,gJg) B;,2+s+QMdx&rSZbJDtz`~ЃN0 NPN0 ч:(I#L5(cv\@ϰ z^ϛi}Jbtcפ@ 8 Ț!jȋ yB=E i;65o>]V*SlL6%Qhb]]1LKTbs,Iֽr'_k6130ɪ :G$QN:e}gKmnV V.Tm=w71cqSu-&SiI]3ZHb*_}!JF?: h6km#Dz/t9|?LһiLwϠӍv'05%G8{Y-%RV 'V,^Kk[GQE]^N:R'qapvЄ.0;m3{'fndD̼bHP@G|8 !zAh^ҖJ#x `7[e^hRm̊BK7wLxj@B ; t'3u1K#N[%w6..I&Džu  T`i8.?q6ydf85c2(,1@\ܙ9=^ x2c׾Y)s:z;1Kc^&狖K4fI$:` ")3t2&UHmlg]̻>bOλLGY{a#*hL0f=RQ M{'K2&DOHė[@NP:'=.*\[/, !fU͜YItHvej $:(CbsRxlCdn%Iڐr95W[Dr՛%\CDe\K;ى\,o(m)HݮRLj(CB[ 18yJP-c_̜5&w!H))SL4pO"KFD^s+$I= ʒ` Cy6IT6>P\Tg0fft\HVtrc:F91X@`i 1\rzEd !BG^Ec}l|l!QW};Ihk8 mp`P|p jI?ʍyTrLP38 I%Q+/knpP|uP-z;7})/p:\mg;7?3WGlv;k ٭Mz-!l8G,SebCxgՄ`<h:x,Ti "MN*r=XsT (,bc1F{[SOk !]\qSLTsP,ivwQߞWE*\[}]u77=Xiũ2n[nF藋<ߜѼq0z%i) KkIB_i[1[Yޣ#F1 [awq<^5 =.sp:8_R9MLk,/d[muqg5lC? F" N;^2so޿{Ο!pjڄ߶2- SE뮊򚢹tYYS->B˭ ʁ8Pr(L$]X]gZ"6_)5`Z#(WV B b̀@gfYYK78_`J1[S8v(рW+P-m𣶅HgH`<%r{`MO~=yD[3 VTi^:(5QQpPWg$$zRёf|.+W=Wu"j wȀ* ;HEiqzJ+]ЁT"SY1"ENJ巳>EG*f)yl|"췇.=qnu.xEOy"vX)JG)Y\.zLgΘϻ?y6*I,%Z$q뵑$ZTΆ@pVH$nuƐA S21C7D/(μTjhu-.NU.17Go3Zpe6$^JVc3'bRZ;tt<UQT 4# Qw1!GKp*ޖI`^C ysvo>M(% I! RF}:%*EKZm>in{v(>'-|gw{&FX֗aʴ %pj\SUXa>8:&xͧ~„}*2kEn} (vU$x*U}%%:;OJ{{yg,jSV_o޲"zjPZ=7Oo,)=vohz!YnpF_?%I(W) .KA,s 460LpJ)Rʞ*pf2Z(n! TH)%a&qMrj'"o%U$ʍ#t0B?D9dyT?vj=|*VOA ʞ8O̽{q4?ZΟ&Oìy4if}M_=%\pcPᆕxU>HIY ARkJ9(b}oW$7IfbH2C7e]0AA(-2EpD3-g`7Cz9ѮD[=bL|2ۡs`r:IM^XEYL!/2,UI4t{bYmYGz@d2h`4/cD3.1*/ךYp\騼z/HގdV!/IZ/A L%,/'|#2:Ř'x2&,* F꘬gA4) )T pzd̜#cwXXHc Jl/)=_Pp|9Lϟ4MAXvkcObrn>c"\#2Ndݮ6C&{.2½ga+JDl 161aO t2ە9Gl?s_FzǞڳv`[z|RQ?{WFO;mw{; f7;/EYr$ۙ[Ғl%i*VOY"їH[dLLjQ*C̭U0+m| Խ$,CEGtMBVFM#̇Ubj4ѾXFf}X;ܢE#jDY# sK>ld^.$ .2 կ,0$2$<=RVF F=kAXz8?lldI;LV,.o9i+2 IOj)cS꣮z AHӗD%UPUu !)!;DȝflO;"QZ(K:A Bj.c\K " # iJ$Z\ iH%f6$LZ'$ ^rV td\ϖuol9 >Vo(Qw/?}(Ѝ[գy|NJ,uf0x6 ]+T0^A`I׍ @HU@fiu%Ma^EciGaEjyW}X]v\[|W*TͬҒ@FY].}#RC2G^IR匒U{K(HNR}d4z!C 3l/}e5;z)J($~Ӌ)$^Þ6~#Ύ^S~WWlEbPJ#%,|cC''b٭l\ΦH}oS㻖cU)B6CN\rk5 6ߴNfnum,ʸ:[xqރbPXʢ1*d!*+TAF (}φd^~g@<90ki&x|Ȩҙt0 e> !l}y%uhs}֛!Axl!hN#>7!$0h&H2(Ap ;넣GPxeA1 CDbiΌ-$%2sX`,bٱ@[xdd +lI 0.|z+QA+5W/$Yx("q$ރV,-!E!pRsCMow]ӢCf?01'E}{va3԰^Cn=^TiWߺOATt'eHwdD-0M^[)s%j|9":Y!|\lbVA䃳*1h.Tiя>NSAWY_&I+i%iEfX|~jjo7Ǘgh5y;6ngӋt~rȸ:pzǥ$Og8oM )yI8nI/@m<ZI76T>:5f#եrvi:uIh~~UW=yoת]f=ʒf]4ا5F%sz~uvV=K:Y /qS'U7pL6ߚ:?-\'ju 86YV®?HIVѰ쭉[~A9~GIm?+懒ۼ;OengcM<7f׽uw8"{S2gE!#ڙF {gFjv!C. kmrt +K$wͳl_4f O`?Spw @RZQuZ 4l:2|SU`ۃ۱{gӯ[%x/"Wu vz"XYŨ/ԗ _?on/FebsP9ⷺ?=\XyrA<ۓ?n׻βv2o4>m-)q9/CKY31sLBC\PW6N& }m/]05 LRq옠fހڤ턷y&g.1ұfELɼo|%b[#D=w "tg3, Ͳ0^D\hq996B\f) , ^NDɿϞzzK3@III1\cQ"ɔ#sm׫ |ZP{])Sܯ]x+ƕ_w 0Ǽ~ =ʞj^@;8TIr.ʣjeS.vـv 3'ԉhA-l)]Qˤ~+~?Wlo܉usB]Ng踐DNdV{Sa"]/rYUojT[A}TTi_(VBmX~TgtF[o5`U#fW_^jm-\[L(nC:;΋pXŒ\($靤(2L$_(ܒ!l! A} YJd9/uZw*hL퍜-!tJ ~ʝun ]]Wȵ5 uե/泽Nd1o}"7=f&m,9rb6jAϷ[r=?̦5-ndE-Iwij|C-@mɞWA^Ϛ_4]ݴǏM;އnZ|s2mgI>}ꮪşΎ=o5oӧY2$"y*j2 QH໿Z@u{p; N&2ˢv٣IJd~/# Jިq\(QK>`u'|x$x`'SfS寯ZPmlsR 3oDYK1'&ӥ8ɸI^%8 h@ʥ^ _!dU“P|9%Q;W+_sܐS4d_1x/Y:+U,o,w<n$r6!<ÿݗȼ^l=U$W1WIl|2zYH}HGίFדZN[^9>waP$X1.$kRj%_|3"p3GL΃VŢș(Y&_yz烻4,J}wRb0|\oPW&K*iGLs&Gmѻ<K@3 Gh!@#?E>_8±wf%qV.'G 5)"<%.MiBuE8k~L}4 ~?ѥI;K4o?՚_]moG+}IrR"8866VU"Rr,_u I2m@3~hV}$57]]0dc%RF+.` +%ΜV7˛Yl3X\OA/ʵz3"df0Y~`5KH/i< (aQe4㇃m阽IG%h$Fm+d=8Na >?G.Ezw8X"dⷒXoP^.\_:9{}o}s&㳓7/aV:8,GQ h -ƫ"欫_g\]>OmM֬@n>WqYJRcչd,Cz{mWln7.\EbK^i9]lB9ä:̒ܣJ6pdց%8VZnX14Q[ϨarD}u9nc(j <ԕг(L.RARTY)q{(ۿiNuiM&6j39RWFV=Z1= i9G;6RQŏe8+(eR2Ƀ+fgzW/l )t[C1pi[ae!g+NE9 ΤKT(?[@̛Wm+ گ>2^ H޺>6`!qɘcc) II}4`#Q R-V,BJao>apD )jX 8(,<5GŰ1 ،h14c4PdVvL󞥎]{X{)));xG`1li0=^c;-d#8O]w*-mܱS3GMN5I >6g_F=q$wbÝWy {,TSl#aC@|i G~ozM|u. ,2Z+$#! ^D,TQ%'d-O+'ܟϖ?@#&hp~fk=(zFi)b` ubcKnT`Pmgcxi~!z׵hu6 bANQۑ]Ϯfn->-KmM$Bk7RaM8%Sz<* ~XB{f|3QpL.\DagU(]VoV# ũbq0^\2oOWW4D㮇۹ODvMLާ7MW+Wrw~|)oAn"rTHNsYiPB襊`%6DF-?7+$__gK-p=óFΥ7MqĴw6h5!#dOBGcLx7z$.wv 5Mmע_jgfaӽ@)1C >MAY&8,gU'͟H9yؑ |{aRCwN2= LW ʼnf!w`YYG(:&hw {-@T'R$]-AlKdh GBL9 ^AxL:%c)ky:nܡ!cFYp{iU"bK8aFhNPƖY[_yIO fs=vDOð/$e_r2I]&Gw}YنʂNE:UcǑt0B uQc,^JSu{~ gs;y,:4;cB! "zረ5\c<р#8`qhfs搤XHMLc1> "' nl~g27zAQ`De*ϽSFɤDDfq*)UB*Fgvcg|kh#v"O"ƞ 偘¤>|!1D0Hz|KC#2\*밊ƥ>CXEdz"|lH%iEM{,ռHOzϼ:4!99^pjD !j:sNS:{Jױұ#5Qֵg[okiq2hL@%INzxηj!牆"Hwc&1ɿbnj N|,"v0Eȼ6X"7K94f"V:Igؙo~>ku KAŸrD  4PR1RjoSq8V9ޘ9D]ԌBU }GUuBhB4F'SF )rl5HIbIYݵ4(z5GhFI#6GEp:*u&MJ{EjW69XEwK'!xXR0#$6l UV-ډ3L4 .M+IN!K%)#Hma /V90k܇tӠ(`5 WWr)9$̞x!{^0YT ob@JN!tP\ J1u]3Kۿ펯^vC_wko%P0 $+:m@)X& oaAz,׆ϒAٌ44$)Ϝ;n+ IJ/*zMe:R\ֺ lQ*U6 ^q[H]RYr<-̩(oOiF3ޗW /o/ghbqi>;(ז$@n!_dqa!0.#1x˦aH063+,'> F `z1{%H5JQI64WFb=88ZG22ϏÑKo`b2⇸aY=gQE zrzW Wo^'/~뛳O0Qg?y 7Yh?M?~e[CxTCSڜuڜrø ԇɸ9ޚͧ7.g۟ 65 3R(9ZNo|\Eb.sŴr.}6! |ְ0iDQ]~*9ydC Gf_oF;O v(G(*JOqm^'_cq4üDQQ!$BϢ$k0LH;I Pg1Ilb;;ҚLl2)n ebǼ:35L 6-@pT!?ZZ(lT: 'ʩ=v76:r( e>'l \j uY!/\ҕ\Z.1q׹2Z!+[V_c;-"޸s+v坣<.5;h8\B1JJ懃^RpC+)`zIr"[M6Ì}ܡ>W{ܡ6pH)B)"Il^YٳgtIr_)RUXsA<>nyq1㓛K6}eVyJ fǮO}8n.GqT"s&r3=g9$_9=8*>ȇh `l0EZz7ZnL)5P0C; 55 5 5ڃ5jXP6,Rg_Zc/픴2#@jL)yoTP@!*eƔ%DQ CMd*0\fr8'ޅ/__BWr4G6szCQwq~vF};{s9yhI^JecN[CYiᝍ!x幋 $# -H8A2Uv%@g 2OgqA#ʁkɳ^hc^Ox}]YS"_>@ &{B_z:Evf7FJۜ괆x86YHቕyz.sc?d}{jl]TfbmSOAn!=D~b B-IdNj S"<'2&霹D ڂZ`K"K{WRa9軒a&pId[i0Y ާ+ 1yǃ[koO>20_$V6vrg+ltݰ3 .Vk礕kvET68kHqu>p }|54oݻ0{qa(wgtD>?3BuΛ/U5jăW<&k7j!. uͼ"]_0uqu'c8 }4ĸ|(o[3bVzcǪĪר Xץpp89Dµl{'%0 ӺN5C#oRHyzk-dBıJm dɸHD#Ymm˶jmAOJ~vPAf $:ˊcnsҖ;i2^rus2wFs,桡ĵJDŽ޼،̡bbO&cJ䃓cώĎ~i8i?H):,F|κFSBhC3L"=j (\d@}yׁAhT`5M(J9g)!G I3B%J?#Gg$VLJs `],KYڌ\e QcBkOhdQVavHݬ uW=pڔ-^Soy43-)(RWuqzܰ-a6K؝6DT`&[gM 74CjS%}VzxhSa[RujN)Joru@ /CP 0b^^hA$CD4]Tv d d/Ө!$IPqCt#Jkb̗~dY 'ŋ(Kl8ȴC"Wdن$!o19xHH ƐeZ3j/aCM6fLȇ5l5d7KhshX jNCn" 14_QD'z}*7^m֩sOQJ~؉x%?Dȭd\@Mɢ<vBGxK!$qNꅷOڶW&׸—[՝t=S=d 32c h8l^(GS,!E2Y7kǖ;2#H)ˉ}uNx=brj&F|ͺӒ&9gu/u,q|.?d:.״]6iMUvú곴f~Kw4N)`|iX)&$%6dחQtKi%iw VH_H`E[z֗ ]AīaJJ{JJ auUL2atUZ 񪿥0ev42 ?ѳf=zGM~UO>'_$0 nLb&.g|?ӤE[iTiRʥ%tNvJd)4EbEkV-8:Ą~h) P&E!/#o>I|N%KrRU"M-VS$cxV/Ѵ)kOyW {5+,veM~7{olɰo^vr@_ 8_h͎N/,xMr#H'N /d|>;5N7i 4\%FthufveH WWJ Mgઈ+MWHWHWE`%;WE\vUppERJ=\FR_ (P(`RKB%不bOZ*,5[\ EzGzt2i:%{jQA|cR ("J[7<,GG8j>| _Cy&(7]q<K%/w:'%~ot>?6ت@Nc{ or+pzV6u!%_$! ]S3DV9ӣV[*J6ŷ^ǃwbדD*'΋(.Ůo7@ۿ ;(?9GqǃgȱdgO|ZᖌmnZ-Kڽv0ny kLdwaE\źBՊ)!=U{=TT"CpU,I\:sH+޳,R.Fp~z8f`68q/\=JZWRGp'^HmU*\qWEZԇWEJk+p!*+*j*>p PJ"\UUR`W ]+ * \iwWEJ#zzpUvl$.ڌ{" cck^iqKB݌l>Q/;=SL4־s5RbOXq NI9D+“#ٮEfl֎5=9r`j _G?`'yϷ@qꅂYŒt_~F˟GG^/ ^5tV5ܔ2. H%=Oô6)y7Hi[s?'r W-TvZO7Ji}44S0,sB#o``[vd,N h|e'na!珟iː,f1 D𺯢֗v˭iyj4LQN2ʀY{'9I Z:`Ftu+@Ls>I t ]} ufS D}C/}*}ܙ+};ܑiaEQ}k,X{v)CNԡxjýi+7=8m%7v8.pfV{qh1h[ ŤfYE"HG~MGO"w<. ~_|r}lƳ|X_=8@JCH3B%DӠK5ȔD-!2EJr":;O~f (Zuk2hnՕ͚4NWX#K'MqĴw6h5!#dNJN=l<Ӡ:),J/I,ooeƊ_mZ}tlwz;̸u͕u$/7oVdU]X5[]d㏷-Mg)x>䩓S#4 F93U~`&EJ;x>&KUҒ8/~Ypϥmg@RB{*&YlxMv^|zɫS%oMqG3Vo:U8O> ߱<'=#GIÇ.21fbW1[>WyS6o0&DuV =6-벥j+q]g#9D3AQIH%cZGQ Q+_ 9OsW딜#v$9{?Yӑ+t0)Ca 1,c1$\S W^\GCJ9ō`ZEĖ_zf9A 3["hֲϪϵx ҧzRu'fF1z `*{Ĝ2R5qWU5}2={y)f\Fu2`eA qD:o4Ǥ:x`^ 5)` =t)]kb[p:Js xu*CVN֚ kt1)X:4 hFh޵󥹝_5y궉qϪr-2Z%4Hԑ($F{笰xD91ɺ :At`(3`qgт5#*$;AJ!ygcg ~g}x:cOmr1mMQْk슓eB&[# t﹖%vI=QJRZHRJ}ey8 s~vG9]*sst`) E N&xs\lnt;~Y@\Fa'F q2nar}Hy%l+>: y|z %(02x0O>e.I{{'٣+EX`H)D0 \JA)n`fpmG t|_[C[DOY}6浫N]pq·_)5uҜp G+KܵT Ȗ5]ӻU! C7n(YȬ%8@}+-7a g3*4F>`ؽ"Q`d(}nCj}lGA0/%QԀUy))gQ5\& ,g1It`2f2.3e?9;_GTL21L`ɣWoG◣q%b`JB( 2 b4ުS2%QG G.e~K ( [:n$ZGj+G^ "R, $sap7XD!e!x0`Y1c"^ˈiDk4 cDdƈuאgԱN)EUvۼEۇ\]`䕉*F Qbi!5B>x6\y,qɛ1`k U cW逰 sVkZSJx0<}-jTX6 r]}S_?ǓqO4F~Xv ۇܙ|4kxж+`I+97r }} \y.‡Ǭ .b?Az)Mk4ܧ ZX@tJ!aK/lIxym0]J Et9>XJŃ2_`V`\^%TOk!:6x0X#KugşǕ|G92or쵲2aDD&p TQ͍QmQ'(C$_8>f`{J>_hO& 0aCs9!^3Q0/?ʑ8Zr&RjKf(էM0&OW^ŷx7uP N|=b8B8a4㸳Ѯ`Ǜ iK@ZMlE̋ZO'% [pO3>&fmSI@w→?{Ƒ]GGU?wpŢ4);~Cd=3$"Cη?]SU 7UKqA/yד/; VB<7c"L&i7rM-U{oTLV$e,4Me=9▜oqr>N^-zvK2O% ϢCod;W/򲼺齛ٗf ݴ|fRm0;?OpQ3h:9`G6iu7c&yt@82,^NƟ.Jj'VSӁ9WO}a0UvRJAd+*dž}Rg 1-!f[1ige6 PyY*l!Kbs䱬ۨPˌb1R Rd?sy^yunn+pr"z#(/m : rcIF>Z)!G4֧6m:n*[Ө?R"6/s@nʒ9ZccTʘC0 / ,gPqNZL[QqE# ې0br" B4"5kLh:SSrNCn)wvEcvU^+ڔ^›'jspR雫eC(N6 BՏnHLQLq@ @r%#u9=+ 114| R0z]&gpHq%.&ccr&csJc,l2d` ,|Vz?*wo{e*7?/Ny`UMuR}f~XVrWx3)㴊|@j:@{. &RlJ ja1cH|)Mӷ:4xWB̶P1h {Y|P9bc0" }H8:ǽUhc ȬLaF]@2y(2ibML 9gnj Lzg6&iA^ƶ1ؖa"=ܢ$GO4NycJRx&G'nXBXzpƔ6әd rጏ>H*)e2!P$-@K9a"6&g u'.N+['6eu˟a m6ou4{l\@%:5qiRq(Zd7 ar?0.Ձ\ͥW&n6]\ΊM+7B) G[ XU cd;/*Wr}q}ǣ@3?|#!gމ4,zSnEՓ~]|Jo4s\bg|xG:?U]t=J=r9.%dlf۴[;R iɽ8Yy{QnK8i҈&ID4dj߮_䷹Ґo֥b?Yq)2H]ebVTe +uI;? ĩ₡,+*Nnle <aYgO/2n;Cha(/ܧye" ݡWEUzEroy+Rkm{իy_^uW \Vɶ㪨Ԫ/EpEuWE]UQ UQ}A5Ni2UMH.0}X\m.apJ-`\Am+ + CWlL-WEWG+]U]FuWB ZWG+Fû]]UQ**m #@YpEtWE.vƻ*jvT:F\Y+;W6ʮ>s#Qcp{”z=y /x0`xɛ-xl~h9Y|++9\bu57޺h5C.>`ȅάhێR>FL+#u 6pl&x|m WE]5\WE=WO V\tWEL\jf*_;F\cuhw)\\;sJWG+k]Z*;`:]J7{95OՂ иD.2y T-Wv=&[:"Wન5**] $•d:ggpU Λ;J%z\!IcC"VwW1\**`8qej6m; %;Ńco72 K&(lg0] ZڎRGiq5\fune^@$F%pu  <҃%D0]&S~iQP2r^=٩aa[ޤ<=uoog@YN X'LB"j0}qyI-*\E}UsWC 0[y Nf'noə5Zv2Y)zmd(ymTܩDɷgWmfWପ*]Azʾgwu\ewU5aE1K{Bll$L]v'.jU+xBlͅ덅]qUk\37nHfST~E(qjV^õ[ۨIbeSҳ(%SJ2!>TKZz͛l8SSZ+ZQ[%4wՎUp;+,UwpUuN"uqE*a+f׫S&j Й#7{j/oR|Wն]υB:l/YgpU5 W&mUQ JHH\BwWE-㪨TJJBuWE\ݙ`5UQi YK~\`vUxW WE%W=W֮:+Nsiϑ*8sҚ]lLg0]ZLZ!.*%>BL+ީ5"6+fWo;J#4(Ż`"ךZ5re qu<2ǮUls=؍Zz直";+# H8G*JՋ^)C}+{J^6Rk#lJo+j/H0*\Zc[+kWǀ׷s+.#i'+- Ǽ|$_I,Bfw/,v^ods&ԴAbYǜ{@)ۣV14Ÿ:*s8tӈpy熱t7qj:OXj%RuӖH P^r.\̾2 4J"_=_Ysզ:uY(/`TJwh'!T {ϛ;!hUU oZϩTRnXm5ljxIl1@جc;Skn`vs=i:wK9j{i g̵16cAFUyڄҙܔ`/DmT/ *Z}%~tvi"%h'X'4/V[ZQW!5*dlzaCKU+ܜ-:LAZ 7."O֮ TG)}b= eDA!vAy1i"! 5y]%@_!8f;@]Ii( "<ѓf W nk 9%bkVcCX DA5DC7HX] @T3be@0 9<,(D)يG+۰,ALJ3-VUq2$?< B#jw7ygjތXT&N*b,zP>qVH blSN,UhE4?Vstҙtv4YTd52 ZzkJB 2ߤn$l& -/l5S|cȵ6vٗ`znzz}—bM{}w2Ɏ, BLtsYhl) l\c F> U-iQi6fY#e0-6n(g3޽.2#lRӑ{8ݰ)QC^"$u-ҐQUr#a?EtnJdv `%.(H H,@z|!(!=:T}f=. >i`E^1H"^S:Ԇ&@u#'mP ˪ Q GʅPeQERPGq1*1<RuJ` LB)1QFXǤz54 hc;(ڭC5iփ*H]6 |tg҃d &S@rZx tmKzzAцp57לAPK}4 bV9ڡbwU!MZSDàe ̀|12 =pe6|_HmhJ7a#5>Nh&= f66PCX1:a%ɀn [En1[.ՂʠM4& XTdǢhBM:SjŇPpͻ %:)`Z-, \g'b!(E%蕲`j?&]ջeZ9,TVXGXrvY1(-ߩ7(N_<}bV7_RPnKRCM/ߟ|\b卤9`X k?j|q3ӷ/[/_oN wЛ/g׸ҏ'O |pjV|~W3Jg\{?k^]`ADYPX4Z;:u%FK`\zoM0Sp+u4%B?:]eQҕ |FtŀgDW w>pFӕ J:Bc nN 'fCW ׺uOWr+c ]]!O[Puw|K>q=[|ZܴoFx~}:4KC Μ=SvgO_WGfs;naݙJgڞ:V۟>nmVb3P ulez^{ /_]SMu3yG>yjƚ_.ۯ1]oethTri=N2dg؋Wodz|Wz2rOnlZS?_e${ʄeqK|u}[nڷ{JF_wčNKYvMcIUeq%-SdJ!qJ_hruFmK\XuK>(du p]lYe|byrZ%9 sݭ,[ i,)'L%?!V QxQ:טڗ}(HxP:K7bw-] GvВUmŠRd?v iF3pi6FVwYZ Z~ +*Dl p ]1}Wt[]pPz6tp ]1Z{%)cӗ+l&5Zc:!tuŷ\Gqj/޻ǡP\1(tKSV͈O+%mBWքC+FIŽ +Kf.th:tbAÌy$)s+FK_ 2JZ[pәӧ1v}ƽ|h* M3p4(1ҴWa;C>]y]1\Ks+F V*hzv 8MMNWҊ:Fj5 l0b:]1脮vJW 8̧dq6t/oV&? ]{.}:U:c7}~h#ec954{>Dfe_Cg♽V,ʇ}v:}{7Hl5{-rZy.oڃ𠀽ӌ0\oZ`[ @UZ"o9c1Άns+ :]1J#tutH{5'u_'Sõ9hgʨiB W?ՋK#e o_j:V0P6ͷ?yqAߟ,_Cǿt+f}p.զiɇ:}?_OE?o_dy:6f]zo |l*z񛁜<98 ί=dHDe@0݅ӆg,a[(:8DD }~uoOr{+WW/L%wt;63)^vtꃷ1O c?p7xgO_ޢlkmvF;pGfݯKGN9I5v>%i.R hI?Q(ZR]%ZÝO# h҇NW2j? ]ғ:UNA>6]8Gn 7>mƽjHjOqWmz~Zu b~(X0S<:r)awqqJsUBX'#@D=q%jUrrWL şvk͟bZ;^*aw9+|`y K*\ʮ3_T$TSXVx>/[bnurT.pXN] %*D,בֿXBPi|wFTjuZ/JIPƃħQwlsDo/lk%Oep'f-*dh&z2}%}HƞfG1hf#qV~5}֐dҫQ8Hx!pNUY f+|7~d2z?x]@3Kq9n[7 ~>j9fvBImHz^GMWEbW6}â]}a5}y! |6̋jhd[?t<Xv7Zj\9:[0VVkE/6ͳqL:5AŴVlT晱2XT3АRJ3?_]Ossϥ>jʐpqm$Q+qɘcc)W!nkG.Tϙ A"߸1ZI`ņ;#!X< FÆZ'ǁ -(ӏOgC= y9+q$drOYhBV(A"yn(Y#$D5>A+%Ǔz`HR,v@`ܨޡ`-1( !F{czU]> mCVJQߖ]n#ɞ>xbMАc1̾u1~N,Sٗ´뽧%=>8;գy9bFۺ/Q5 _4`U4EԘ}BB*EPU8^Y|j!4]7AtPz*)*xD-@AnZ.u=K:Aʝ9)']Dk  ߂X~ȧ wH@ѸxzYD`qU|𹊄jչFI x.F] ;bN^-_}s@U-$+oi`씄};L0H8hsPy5ꢷ 7 40ehmZ6.ti m`5[]rU_o4Ea,;_3F>|vY>ffD Z-vo"]ECތy]B45xIջp\β=H86e9 {=JMl^qJ߰0%(|F1BWu7`+CnE>7Ӗ&-4o68NմJ}wkQ޳^j24JNrӬ[~ :VUpL?x}vOpCaFuN``,Tf4㏃jJGm L,]-w`/ J d;f2%,|Z2z UHF*4x~2j:;흝a60fXʹ u)7&FF*#'0{Lc&zϙ}V_%0<ލ3QҊ3˴"3AFW:+8K?@t={H~{;]+t0)UsObYPp, cʽ'\S)_WGCJ9ōaYEĖ m=3 qH ÌМ=Kh~%UVsՉڼ٦) t7&V,?1Yi՛֫ L^Qb#\K_v5;guw v{G#JgRHw؇S ~tkfk{&W"P&zx x>u#mnC(6뇨m] >5-@tFTb{*9u`oLLDT`VǬR E!#3t6Zg?.%6!`lKƨU0 %J9XE ZR5g^  TA?bp6ȓg9ny &8!0J31Ác\a ,(0 ƒ]j} xRH :q`UD4#EOjRqXͷ2~ayg#+/ _<Ɵ9MH`Nc5W9:\4H"hgΩ 9iϤg a#IUyFh߯KJ Π;܁h.L3dH'sMV}O"a)>xd90< O4*k3ƭʬq1 Xc0**zo]:u O9hO^gqf)!k sQdymTEn9r YH :?_\ߚZ@4s|omR`1'B0穢f R{c;ű!0J]mKz}SG$%8DǤ"9Jc;djL$(O =8پrm= g5̺A"ZS1xfhjp##<Qpjo6k yPDa9'@XԸ6mv`~.)yt"(OAV %hH)C^A:RJѦΖ:, LO9$t>.ſW<.@j&;GP%)Xs)":yDl0\B dظSԩh+r$`5n\ږZzQN׌/#a@#MљbLYs;<*g ϏHgM[(؀NXVXVO- r!QQ(fMt:?~`E\, IJ߷x$:9Y72Hb)Y ,eST]5i4b[H~XR0y:2NEknjx1 /W5hdqnEgzm^튣HyyqrhB=Cc$P]:Gb|H]ÐafX#P++XW"4~\.ۅ.cl$Ryv]v͕QBb2cG|@>?. #< 7Xhj`yQ7\>9Oo?z_Ϸ/>?x& V`l G({_/вavh)B}κY]e}Nc5bʠbo// ϼCIauUarkY;Frҋ˫z!qB.BTt>AqW6g%b`2BȔ1 'ʩC)t,Y(HrȿB)TRkPQJőׂTKf4-!\r"n<,D!e!x0`Y1cE佖豉h ѳ:;>lާ Z`sWցŎ},c Tb ]mo+?E';/nEAvFIH%G#2eIwlV3\^syyO{ȹ;w7GQpʀxe:HX4A` TЬܣ RS9 ُ΁*ݙ`; d1q .Dx=JŬ`QܽXw#g}s5~;"PKߵiۡUoh$#8=Vd'%u2@;mFzAgIi ; K-[.oYRC6].gZ3&UĖ.X;GJq+AɎ9nPS*+*<DkUϿm&hRHTAF5\9, U&&?w(lTrMD%{rR=2/Fkrxk/'KS܉l/ mye?={0}v 8DžB:*&aJ sH(#7:MHܫ rPŽW{7#S-( g[/0.YG |?B:3S i8֨ؐlSL] 6 e.UՍ@}tu (^лS-!;g^zWKn ^0zc+N@f8_>}-D߭^hַmEZBZ i=ڭ7ROR B$Xy ql L QivE[ ;.wr'ed<dRFG|ЊIADmw44ș10~8E# W3*)oR^g"z uHƱ`c)T Iź3*|a12R +N32~KӃ[?=9p4M^OkAc1 a8!N*h2jQ́'!=1βJ{%A3tBnG$&L'S.챋ugvAXⵋYǑxmF׮v{c* TNL;[Q_ Rۯ/8k|~0Pɧ={َN*./O0{!s(>9ނoo{1KE gpYKEg-X]>}gog%!Ð?Gt.%U#tvraӉe^o[(Z0vnS{y: j~7xz4OA~M^l~^Bs ,5 Gw] o/z34RO ^s{~ag*; XuJ| %OsY>oSqF2m.@jjީ/(D.IP\q"3\Qi{r( YʧpNw/pᲕPYR>l"oh2\2R3чIrBre8FՎJ,YAgqlw=Y3Ts qp҆]M3_>Y8Y_[,mm;RCm 2m䦢?'+Q 5:֐:F sU.V]X=_bb)ז>(΅#$k7q<*IJŤDH助& H藥gSGnO<,Ǵdź9@ Sb[d>x3F˼l[S y"#5+^= Vs裉=(W܌~r@PLRp 73(JƿE#CAU DD3whj Ad=ˤr5 r-DU"Gyu%6\/9%e?>7_-&ڪhB.^@MU`$4ñ";OJ ]o;Co/Dre4mMǂ)Ϸ];ͬf .yWooALU߄uA OI *tpJeA @bT;c){'x@ > @1$L ã&qM*" TXT5]$ʍYDŽawkM\?ƺkקkYOBALe';=ϹJ[#}3R0^>{( Uϲ)Lfe"{v ǯ62mroq ϼ7LWx'rЉ9\>@3?؉#yGqA[cM>Ro%MH&%x4${TܻLFNRe2P3gy۝l0^KVkϠ=sHYFKQj Ɗ 9IC\PNJTwNJDѿt4??o T~tn|;tJxc/S Gp#.QI M߀,z䃜Qo6##d[qu 11&͈ 2m ʢHũBҊ.V{#GfH_lY[Z;q=£59ZMCXAɩKlZ) I!F&NS=ϋ<,\Bms#Ǡ!dG\R'hlnz  Zp9N~@O܂*E$wcZ$ɉ0<1+K2]Q_lТcn?J 6uQ4t;'NK%xm4'/܍s$>Wt\rA[cM>9KN̈́dR&~w)B:JMIع4wP<_' e*"9Qν)ݾ^on諹̄/ m(~ؼ߃Ť$UFCNjQ\u9}~h%|- _u F1L%/7W7 5tMeM^;T+K7eM=6QId72H%'їVM^*dxuզP] iv 䆤1rmg]߮]jku6 OhO?[j_ _dv'blNzSr{sv7ϝ7ՏxG]6oW4(?/-"ژ}-$r*{M+懜zw"k9rA):_ԏ_>Zzפ(:H1ܣ!t^D pz$4 4e{vvuZS#elw`T! M ^+2Yʷ_)Y)!Q ,Q^:^),!)4JTMIřAp =h &64C¢::q0 Xy zJut ^". t}Ұj3.;jbo|EXPIDBlVEq !jR8":X8HV}Wl[(k~V6n0#0p_)q~q1dfSW:*CV M7m/?M`'o/w8L\/8<~(R8uQ!S#* E"ZgՆga}imI|̰Lx!o"Em%c+t1oz<(m @ENYlmw\Qtu; ?nqSLoǣI-|O?-QPmtlXn\3ooOLs {3%߮ X) 1Jg}*#eN-YEsxVjv^%οDUPΘS&|"X]X!$-HQ3bH@9I1@罺2>k>jq[|emSq\bկ#1WJR6ZHjKbZCs<*:iyS{:["6gQ\b ^scmNj85mmM{nel2lcJNS:۔MTVC@3+BKJL ,OLԦm624$" C|.SJu4yn/!)oҸnQƔ\Ή/f.!O*B(U#DIZ sb$vSmx!r 0nA~{% '40Y?paejs#:[D.ⴸR٪b㈒sCy96|Tߚ\w*SWINUS> xLfs$ T-0J('䆜\h9N& N&SW'qhٽAhά䍣,;$jMY=,;?\GyѴJEէچ [`xWJF GoU޷ve5qNVOakӄa8P'y6^q[HSjX~<_+sݯՍمer#_4\\c.p~͂|tqviG;AȚmI[Z [ ,OC]W2 Gghx7}tC42y'׵3fNrB.ڭځ_S`R̳TUŮYG|zmG l\W`R+IXFߣW!b7"XzG֦wlnSGN^bDfEpHvO)d. %!Rܽp \ 8e^Y-1O61ЧIk"MެϤCЗ5^7ߤLHɛͭD~>] ge܀.uGR8!JRbMMuwhnZV);$uQD0y!My>&HKHE*'tҙ212c`>`n!6, k'. vviOM@@\`V <a ܡr"锨E0bJn h0*w>7ZnŷrwBq]}b$&PIyuB2)>7/'X9q+ه}}yteƓ-c̟|{+9}LWn;9mxMԵ{ow3¶dJf_oE.g`wlTHTOXʢװ 5XWK9"UzPD Wrspv(wxG:5XF+|fdx,`9lm0s\Zs40h~6!VJ7{<BYS0 SdcSF nunA Up[M%Iǧsg[X9iU#0Z Fy8rZUiះV㤅JhYZxO I:Z؋;-/u.~WdVMN䜻RXJcC*gB]Nů:``4Y(DH> F:Cdց֚9Z[3f5Np8n9F~C)ȃ1Ed W%2b`[(DBs["KH¤5'lkW h#]+vd L.%lRHD!$Ġ%P+!(pSMmօ`4pd/l[=Λ!y'3ɨ7=</W9 =ɌVܻ{vtvD\ulΘ3\)b6hm6gz+43`B+,)pYgN2Z.2JџFq[萸ʀvF\eFwE\!ZshUFe/^FvI\UB f/l%2ZzqQj+W0?mxKXΦr93מgKN(U0V2Qy*)ſ~4h~#KJK* B2~"3TpbYuj; 8:T'8Aje5Zף'٬J(!t|3Lo~~bM47X!LNQN(VzY3?σxfيNBs1|-wL,RpJ(=տr٫}Uw &7%FG݋W|O/ҶJUũ73<%4'e{x2R I9Zd/r?.ϥF cO3#͘s4LQe8+O=O0?%}1R+ŏDb出@#Wx`i7?z?lUwRv}fpkTy}B`N1*h(Wym W,RGZpue1K-$Lܯlm+קw1o^^ e-`_, <~J~y5<ы??ZS)N֋U?>5TW|370_[mc6GPޫ*ͤ5wA 1ˏN—uXᤧ@0>rYcu Χ6,w-kޫ~@=}=~+#JޕTO[&vS巗IǢ c#F1-Vs Xu*JX PaR/1",DDꥦ0+CʘtDž:Cgw%d`~)*VGÊ19lQk/ bֹ+ͥ87v:_]^뮊ֽ)Q+qɘcciW$$Vj 6h%I$B劌(A0$M bXRXdF a J/: X& ydcR2E4^JJ@pqJ7bi1(.QX G: GM[QlR5ܴ9hnUmVLv52u|&0|ȿsc wGC \ -8x`AN# N[Q3/(zZEQ|w* ,2#W 9HH=9"Q# +(@FɓuƉ/*!q ">jg<Gj"f9^"6 F@kFiq6:\7ޮ~3-<х␔]HK)ڲBǴ@#} !t4d} \?6 (/Ƌ7a~9Z>nĥu+Fţ:oޅIL_׸_ǭN#wS ħ VSX7?LA-D9/$"wj!4W Hʘ-_N%"AJT r.f.i)Zg݆wD>j  ߃X6Of<5g EM [Lާ"jn$P$@\ lf "C޵b/m V= ]W/Lz Xf;%P9D 0A=֨  \M<@ۨ[e:ne zm6Eg3f+Q6UJVS߰!wlUq'ggONONYYYp-;}|j$錩/()FxXZ}U MQWc_\fT+OLfL%w@ʫ%>W~nI/ yjC3Ozt*QbC5:_YdBo]B_3TcA Č;_~ف)`ǥJui-nKM޺{7轝%b1SR4jij|u1ΛAҎM3'Ϡ#=h>*ȷy]wN̽X &J1qԜu߱Zp iZ))qTZV٠ =1ש1c3544s#y3+Ty?_~M]i O j Aed\ h7)0?oޞUu:+1$+e3hy|4b'>n^G.v:??NvV?~𔝎;KJ&HPAeJfA *jB0RHMT"Z[NނIpώb:z_.`MQ}LD^蔽5mN;yZ<:+bsɌ;6PzgQ1B 33 BOD% t"jHי3X󶢽͛w.^ >6~ޢcelh5աYNnAVt+p_Ti)`%,]=H~{iӔ̶dKd3i.44~1·9/3"ÅD3AQFHceZG Q+_=fx Յc]'4 \ @0v@rKrou )kyѭ!cFYаiU"bKRm=3 qH ÌМmB.שZUkg݉ln 6V,ߎ|Jޮ+0ҺW5&L˫14?!O:PS23Q"6ujԀ \jmEj+G^ "Kf4![mFd&&(P!x0k532b=6MVHKDjjg왆͇t hMb$&s׻jk};"0ʨ"$ 3i1tu z:[0ȒK3[Q+%척Iw{)"Ѐ.1ۂؔ;+{w_LjމٹMۖU.jȍa>nWd~Nn_c ބxቍYVҬ_w|=j}C7PG# #]n7+鐌v͟6'\Kٽm_}%>O\Fu`c1XfX"(#mD]vaBNhFgS7*5)BĎR~Vnmc *`2nxQKniVh]w>3Way5[u=z߁- UJ1LBkw G#hSN9؁`@4NI. 62 d_lrD`p'93tvsF=;d r˶w$)nW-,UWMkTtf*&l9?J Fc%ޗSWq现[ i:°H:!2FΌ4Hᥔ\vGӹCv~B<]\9cB! "zረ5\c\sƞ@am a䐤`p `S-yA)q*cb ޮmViz2!frea5NƄL=lBǢk{Z7z[^%D Y٫W$9k^Q71HY>NtޏYr[ Sw$ 7G;2&+Oqm9F"ҟb&å~j*a@Yi J!]rڌ.h~j"}rq(L"ORNXo>ӏ73 V Vu[D*[(Ì$67}W/ LRʊ+8+pJ*(ɳڜ޽{r3¯_eߋZ od4+4G8] ˵z G޵#"<0&Y[Л`e 0xu5z*o\5"~!^x~;;V/j喋j҉?īwIח_۫핐|귿LΟ\HË"ؽE߽p /Mi`^-a;qgU[g _|(r5fϬͮ9n5Gi_Y4_MqqΤBgIߦ !vz W}c[;J!gBV`"LPE!,҈rv,h*vG#Kf)Ndɫt*ڕDc12)G&Ѵc% IZi96tb67^lOzyT5}M1:5LǩaU{CIWrD*NMZ2_Mtpw9 m|!Lh:%[>ַvdKICңktToRsoY.w_>-Z{V1$\xQ d~M"bur*C@6BfM0sʸ)paq<\vg`m3+H߽ozghiT+lWX򓿺^6%Nȫ-O|?6&nk͟puF;ojqHHH8+`H> L0Ȕ.-Ж*#X rBShU60Ow+Zq\wte#|BϓEX\pbqr2 l6(о-`Αy@3[(Eb Qj|=!p?i{5Iޖf^dxё CGMLüG1 ȘUMDez/[P͗L` A%ڊ *8V! &w^>fJYQ*&֖k?މ %Lt)Badq8g >'#QF׸g 6Ik9\OyEe f}*m[Ex;k P^pdewl΂SLE\hdu`_:h+LjwL׾}C x$bUvG/&g PZ5m>S-ti{nDiB>6m l S\brRkH U;#rRQࢷmI\O\_E>6 u.6V]W՝2?' #QLZ =, /3^u7fд~~~>wC!V9750s*D4eϥ^4J #ЈBMB,QEɣYkc!_mhq8V;m։)ib-S߯O{HBs}؈#{eup}RsqZ--k`rqgE!d2XL-qvw]?f1s# 7M SK%t 0c-oK 1%4$R=V:1 AR+/}A A>[3q&N(}I7XLG38Ϸ>5o]|ʉT/Tj H) U5>9cFZV;d.grPb1Yn_=jq@K/,7%*ak\946)h)^V;ZN+DlK8Y M+U$`QQČZԂJnckҥVܿPdVGsIor*d'juT?5>zecC 2SY}MXM}{\R7E\ZZ!db+e"&sxFd0{ M-TBH[J˘VT\dKAOi 2R-c3qJ3[L3v7ƒܮ?Td!>sپ94.0?? b#JD+Y11ք ~Z,1R`ȝdDTۦ:,P=$$j6A1hLFd vܺ"e\x4u8[0Tv38-yn\KR UIL[Q(E01y qT)idâ1ɞO^ce>   I(B.rFh#!:+87 AB#A2(K3!h1묅lc$OZoJH"6ӯ3=@vqbʹdWvŭHdt1Rw-3t X%%/NJP"j_|%G̠c'[iǮZ/vG <4):s=W?J7>j5'!=z?ʌ!IٛןҨA!d A/2ݾ .,(pLI㼎{!iB?9jq:;ҾV 'D8ڐcOE0Y*sV'`\GA~ݛ ڿ/V 7YsqRŧsWun̂tրH* Xuc 2OURgu ?Lk=żJo0g5NkiNtgowX}0]r|z$fFȮ'-:b)C`Efxb(dZUv9ץK!fͅ`3,Z-r10A[dR68IJP|)W!!!כ' Aq3 TXV%EhZ@q+:~?j9;~sswι9;~sswι9;~jwι9;~sswιbsswι9υdG! \eٜIz2DUV)*Y-Bz"-Bz"-Bz"-Bz]ppko[!EHo[tH"d 2ջlSt_7Hn?yN3i-0JsH.{fܙġbx s9 \Xg^Rd@4bE`q e%"gEZ䀕f mN(2PGS:q6%|w}ݗ)gį%4L8CBwoh.Arq]ek0{`gZpiT hOA>2` ϘKO&EKPN/ц_,@ *Z{IlfN e𖩟-9oL囯R|ID:T H5>9cFZ1Jb5x,z{k:^,f3ِ_ӥhcpWbׂwɿ®'lߤb@Όtbr5Lŀb(b5Ĭj}{1ට]-̩rtT˽,<.ikP#GP0 TK)rFR Lߔ=$MYSBvwX)yrǡrMR9|ty$ bn8L8{F`8N3,F`l|O6]y}'|{=&h">L p$N`r[˶ \wݣAT/"dFMJp9BKZ^f"ӭ).Shn y.}3QR*!%Bl!7NK.iR?=.޾9.s0|f? RDzO&7E~.`h<.ۊHi IULk] ϛFPd7vhzwMjӸM8za\:*ӥ7R5di#,~-<[ FEqRy5jb.:T/-Qv_&lt`W)X^IߔlFWЪ5 gϝix\V_>>Ds=Xzl"6%Y $‹fw)*{UǓF5!pI-g:\d~GqtN?Q7I?y!C_ۯOU!ϊii͠,W[n) ){OWZ*Ku}[0uխ!:S=I";*CV_Y :I#YU/&#w韓t62g30CxSʌ iޚq5vFC}_"la Wwq˪{_ #Kʸz?ԫrgqɟKTdn7؟.ۛ2]0;v9HDҎ&=ÃGT DM? 4h6rA}u1|<IrfhQ`,!=%/m" *sJ$Ri`2B%DӠKP!)$&"Z6DFȟ筜N,ܓ|}-< (_FO=!"F8̸iJl4!uRrꙏ`I鵢xOz@ך5X񶢝7U\.54Eu1'P{"[=8x4)XP/LL>Ϡ E9!kEnF pA)#؁2MEKb\N9˜#D!MiSNpM{횸:=V)n z̝vX-"$%3À0 ^ز9ܦ*Wpjv5MYub,7o>¾[t SF6Vh]5`j#d>*xWثyk/-fϒ* ]pIXUAN_xV6%Y |m#,E>zF,>8?@?'aJӛD,#Gm)aT((eE`qYJrV+ٳni>2"c|v PжNBcs-y,Qwж( ^ꐭsu*F2ԥ[Go/zEխٸḫ'i|0AvU!FEWfзԗ;pcv3,&'Ye?\ ~Y[y5H:!rF΍4Hᥔ\o(t8L3&)#~w &;cĠA"GhK'ȋ{#pa"rGqh7CQ{0rHR jㅑJmc}J!Sz&Xk arklUf^T}``WAΘ(;xR_N:k3^?EC*FTb{*9u޺A:I JUS {b$pnץu.%6!`bKƨU0ʨk zqyֲFΎ) 'I9I1X8}ĎZI3<@afcǖ"W8mK &LġA^& 9Vq$\@KeV8e*"Ӂ#EOjRHXͷ qQ\Oouc1XЄ4XsLrlh5"EΜSAr"ӞIς^)N2vD26 .Žu7|L1MV: :^ldϯ-y\0h - pf_%|)5EeElb$f: $M>Eû.=E?'aVAx-yLikc $"aHg:RP4\Mg:%x WmJ`pVtb@)X杰P XË^0,}a,竓zlp0|*QF*y^bU$Ry!fmzVF Y>Feȡ h)i(f#wp6*/lciPE6 *'Q͙/^}U޿&w/a,8:$AhLZ55WOM>5Eͧ|yY0a|>ohnDyˡ?-fOxkxB:2C+U2T?\5ĝULU*D !ZDLovM?4p(Ɓ ,+u` 0V , g3*4F>`ؽ"Q`P뼼$@`z%\eB ر$.@DŘ'WN+;Gߨ5+<8F?xjZ<(QN֐\x}/I:11S.j2t >h<+Ƭr]ҭyoxh.dmϮje՞Mew^n6n WNUk<3sQꑗznhau{-w0\$)e`oUU?pਜfu] xw-~X_S6ҤH /#6l.l;94gӤe^YYrGAcwut]vGc (S MHL1"1sUN{\#E(QkhF,B0`<)V4hA #(H8H[#gG>׀{sas趛5nVxss=,T«}V}<^@sly*iAFjD!5;g#p4Sv󏁞`B4BkM̨׷ۨQ1,%! j@Ccm]Y_'tm_|}qKULQBf\qBO,DԷ~7'}7}>jI5X@"wJ!asͩBOf8H3y{4Sat[4ӈVcL#JOL ?,’Bd†d / gQ[xwWA%pGԈai|FAwt4qVϾ[$8'hn:>Gmon opukx3X;o(ʤL}H~=*Fq//^F(DQϳiN0q/(FD‰ _9S bY @ ϝ_`V\GQ$OE M(ɟu\";@׃Y ֦;D4;%Z$8#z%h|^L*>I-~#Gm.xmbı~/$ /D|榑EXpkdV#rcDI{UWbO$u]лuIic&mmJ3je2ցfgNqD\ (V+3"VMGJ:_miIM`_E_H%Z@RFOAgaS4akX"si\H9L0K'Q?L^6k_-n1g`Al2F%eR((كYI֓r6@f!]K9/ԝlqQGjvKXziYp'3˕wlrPGoyvxmXU\yqTb>2Z灝ɿ{9X&lŒC\{H0a a~ C߾73djګRa;P&9 ש))ӦHxe7SsCzW梟rF Bfy%|Rc9Q"Xȕ"oZ&͍+'(\xJʁ•PM*d/`9W[C[[;J*QQhrŎ &=UtYfG!)YSEtF5.Qmqh>v瀈Rst`Z=נ2#0oop|,)}^パi3{Ӕu/HHC#~0x .gx 믛Н")"\ڲSDu{7v'SĘlBM%M)Q^cZR=є\#2nUjҀ8 q>7ƍp{P~l|mRwr?&ՉޛBtHUaG`?lU՟xJ`#z;|lw$nK12%Dr0a.CGmƼ6X2n9r Yf Y+Bwr .%6A2 ǨU+"FYJy>%&zq5lmGs'\t"{%^~y;NjIA)c<\m'섾Nf*a@YP5 Akj5\鹏L!#Ayqh|IHDr9&#S0(>4==\ΫRH\RI?/\l&gRъWW\R`Ri,iZ:)'_L~,uo~(^\L g+hdqinztQ-Ջ?]EF>  5-Nʖ_ff]LXBa0,x԰nw;|f79V ꬓZ]WFIPw55@>]*1d2:^cRE6T*&ޗF@yۏ_u틏x/^o/?:`k!!vMpzM˦fMSŶ7Mj.kk]|[(=7[D䧫!<}9 lb׬.1'y\w(*b򮹿Sw# \f+@c>e"DQF=#tw}{/(LӌbDb@UE- 5)BĎR~Vfuwˆ^ 32dK EdhA #(H8ɐnXwo NM}lIon$PعGޗ).:BzZOU"MV}7%ָLslyUJ1LɄ$h(G@[1<YayړE =Og"b)JEڒ%ʒ= /j6UUhXt3ĂIb$+UI"4f8quK܋կos5jbJgy RC#je6[4xsp%NK]JuT3~;.wװgT[K8#Qjl!"+b06S2 ѓm(S`m,Z|(30;@m$4!@8u4}9lNłP. d3*QLX;j;V"YW^F)F.7C2hhJbIEA:h 9Yl(} B8@aL%5eDjjq,|HL%3M'p!z 9ZlkǩcKu]odÕUY:%H3$%Ġ;",%G^}ˁg@[7qګ-Z0F5_|j+D k//>=R`^zrQm*]Sܦv+~DBHBGB/|+E $EI c/{jۥ0ܪ07?OT+l|̣Yh$OF|]2?@y1hn{O}~mw~ryғʾ*i>%`nWn*r87Yt7|bAv@ZBZN{CZ7\:qΧΘwrd cb13Mvj}@[ۢ;ƶ"A;f{95E'0`A<6SRKVIx`,F mIo{9g}tL|7bꑀ=G9wS`'j 4Pk!m8#nP$=il<JzUkR֋_jJ/a%*dFcN$ Y2:x*zҔYK"^5{t~aa%jە#,BC9i%2{Ĭ绥Dz趻o~zz^wz;cMR n?J _UܕheU /SNjá7c-uYJ}ʍkb\! Kk/4 7!:U(1*Y>x%yGٛU+q:ƅؿHڙՏqRל.~|U]g7<"^_aO/ /glc5twwٯPq7)NJ|yoU<9z:87ݏ7? +sϙ`䙤B>JC.k˼ok^i) !v^٤"Ұ:&~;j ^d`thkv P>_g*<yܧuM0~OI)m'U] Qvhjemm-< U*F%W8ĠR!')SQ`)7p+c('(JE-]6I<""cttKf_ompң/MydoPN\%zρD+Qޮ<]]|\Vv(l*T !8Ʊ9: 'yR*:@dXMuXYlB p(tH@5ÆFsJ"))YZfAl/jҥ6)5;^00j2 7^a҉.5*/MT$^L.DMx@ވI\CS1XݠMw^ImfKք ๼%Ȼ́e1L>HMV8N&_?d)ײ^20]]'v0ibT_<N_va=115j(hz¸KE{?^>EdbZ{+wBX,Niun%!RFH2|*Aa4)Y@PW&l[2c"Q`D`KQB " (ƿU&Z641l&Ά#r?d_~<^i<+ `]27"j%K&1!TU!Bn%K{"dI#`VF%Z@*1D*kiA%ʥ.yVJzo5,Nb=*>[ޭ}oZyKqIW FZ9lQfCLt:`&BYiKv,ܠiI.P+P Y >6^8Rcxa]Yff~wnwO?^A'Ί@9l8]0&ÀWWsϵ2F㇏-@BCRTTSR],JuXyIm&8>z7ؓ EVUQ}] E%P{GFWA3.ǘ82U=,>>f+@1>;K\_8plzћRWÕM)H*G+t8c!*Q 'CߣWB*A+ )dPba0 0 )¹ģ KTPPStmF VY#;V7٭DXm n>[sMzUͭnmf;ǛYHg^^3љs(,Cds,db4 bP@V(2F OsvqyB5=yuGf]M! FX(Dd"{Gʀ2 FBUCf=:ݫCHLk=+zQ%hX-:Qcf"1`(L$ 7ۅA{uZp`hPOE=Ӹ^@lZ6'i\r3AL5")* ދP ԉ|̌b QT|)泙*[mXe&H±pDNEATd .bs/ i|oyB_WsVoge*9$LgN.򇳫0;1KjVRUl4Ko<«Uޙi*MwtB͚|wV\ۤۨeK5=~jgM ϧs%-?o +nTBMrF.,mFwO nȋ*7LX[NO/Tmmude=Ճ0a &\ZxU.+ٻ8nW%I}l-b)*yhOc؃Yr%"X`h`Ѝ\'ccqR쭧?韓~W30CLwaƿ%בq`˚cɹ@n=_JHƸU|9]\Pն`i`}L8Qݼн蟝,x4hm;YOqUI]/f<Fn4H-/37S\4à~o1c?>5z!ŵ> l1I=7cX X30S,B7\>_F&||cpΛ쯧O5`iEjR[*Լyzo\T픷1z7EjMCmisR6 EX:w 'PrA"t<*%%^ H!)71q"N,ݓ|}>ƮWժ[\x _CvM{i."mRhWi0O#)x+ XhRnnK[W^M7faa(mM'`{8lj+{zσ2^qc>3eäK14sOܽ Z6Zcn8ZģKc)-}$Rx)~׽?8?d䣇̺>N<7[MvD#B%E =Qk0@98GF{-CQ{0rHRIpxaR[ERDdbnlw&w&f7I,C< ӕ2&Iz#LaYtaE[[f|or>-*U$ ^Truv0&1T  ! )tKMR&1j2 Ok NR5c]L-)B)) gQ<{&bB 438D03&] ekBW!6u#2\*밊a,%V6UGQ2:IKc5ߨ@%zUMrzKcAӘc1 pjD !j9DӞIς^)òMKyon]{ 't6*+ź (O8p58yqJk\,`k $ˋaJa\Dp"`(.9r-si̢E*0dt//<= xegQâ MKAŸrD 4PR1RjoSq8V9(D-m^oA:ё9"}L1 s2J9)d㴒= 8D\&(O xkO CLYz@4ʺUPA"Z1\hK Rncc'>EOeF#})c@ S.(93)(4\tF8 tٔ*a0p55guB\wU 4͢0h0Ȭ Z1 UէpUMeHQ}DuN{Č$5+ƎVoA+YkJUsfV]_V&P9L~,ny_O5k.yS|;7N=hŅA[v)@Smv ; -tYtEW6v7U6j4iHS W@O_gk:'Z*A[m;U}ehu!VG*P>;Tu" "E_c\ wXT5 +0N?=?9ӓ~x-: JԣEQh4hZv46M4E^7M.ha|nbmҏoޓlOWbլO bO:E?-\SJ*TW&D tJD^OQ$+uoF;BzF Wd#jLe?++!0/%QԀSygQ5X\& |SJQu:9(ZN>؎h}vϱ}בsEMXSw'u+$OݹUGyu2K*YTF2 b4*2?~qv:82E| (sJ:n$ZGj+G^ "!hKC̃E4܊N1"xD!e!x0Y1c"^ˈiDk4 cDt;gKߧ͇t箜䪊krvS%6,ԽQEk:}c3Q4솪UxGYUgCin%Z;3t$ T8]b]/Yw{AK.;a;y.Wnx|rׂ5/~>LF5Phs~43n؛pQ'e5{ekv}uۀjml:g*u&{uH}g%QvowWc>I;Qzexؗ2V KDew |rBG4=SH k\#E(Q*ѭN BXYL^j("b1h#2&"ݱޙ8[k\9>_} ; PXVu׶CM|in۪&OwT z9{8yߎ+,`bc Cӂ#4[pbab&2TAU |ke:i녵T!$UHQ+NBT P93E0RBqK#"q=Pάtvl4gYY,N )2)䳝p^hap$RB!g LOBgTX :?+/lh2;MC;¸ۊׄgr`ysʺ?~7gC⾫g1bagaV~tV$G%ԖhQjc7ߠ2g{}:)`u81mmFy1VL3z/Nů>mhGvj.pU vY4Vs2l9H2gkn{ XeW⪇rXP"z|[h>𕍷0,f^Z*/aEEq0ȕ^W}si8@IXO{5|X/ت@ANau^Zmw[YhKvW.i=LR7ȪVڕbhKҹXxiJۡ3Ń5bF[%rM]z>{ß<2(gyeF07@Bj|+lBYt~7l6HyF9d(grE:jM+lNp@ME @mDF7R6.=:B@L?# E գnA*LO.N8q~}wJK'%[:.fdQ,NcXYg_^^i<6 -o,iᄱ> m^V[{PzM〲,mhuzŻQו_ԕY6~W|2Ʒ '~(\y}5Eevڑy4OHV/~`iqGGN,'N$;E#S~Z{Y~39տ7};|vedz/'?vzmڽ9\jRzHpY%$/s#anAX1s"Rj1L /k c~%a<+ (@' 5U3sVEX!;RMf]7o{ 3߮I޳$s߿4ZP׷mS>b達)Jh_+/r?fkzс4"9JQ3&N׿}ѽxMv>nd )Clp,yo;иʡ־t09x;3o~ek Ej^z5%B!X2!$O~|7ܝAxD5Џw$́$ ^24P}@ْJ1 YxK:|YcIؼ}ƻEBicpoYJ'Ju Je3h"39@?L 2h'e+\&e;#UcIN[,V0)glʭ:ւ&PСʨlT%'5 7 f$1Kyk/[oͿvg L ;2_r+pѢz[/ٻ{z H&3̮ LA]ՆIZZj!2y䞮fan|1ig\j'3uK=hh3c*(Zї =++* kۿ2.w]nFhq _H|Ohμi1 7Yf^QcZJkr֭C^0i~N+\Lw`SԢ$;"!Q1]lj:}i봊)R4*mBdfTH)l" W1{1MNvePVN1ĜBn(]en[0 5G@ӻJ7sv/dn߿.|JB`tP+ؚHeNy#eDЮRM)vy-7js|R㧋Kr.rDuEW[|CS'*-"Ē}-9,cWd4֜kUQ3J ,%G6Wu%Z1EFJ 02*ݰg숅ZvB?a^ʢvm3rzrypŷEZX^ lvz1;YpĮVՄ޴֦9ds"zPDUe952ITu8ȷ5`l2 f}tj􅍯T}˜; jw󎁠&ԞDŽfe⩏[IQ*ڔ3%*) hꊇ )#+df YE'(&vK#c?bebQ]͆OnPG: \l.kyv:"2N$>m "UeM b-YN{ rp%p!w d rHfUne5Ht QT;[B&dzolN*z{'oy~Yk:E1UbE<(EV02E-\7eӟ: Wky$׬Lo/k^}չy 5R>NL@U9rD}ʛ4>\3>Z7>J?)}ʡuR#f1fñh%:\5+jD #+6dWZWJ\B"INN v4ڕrsf~ +d0H;wCe/Gg竗f=$T#%i~Uաo֌`x 0n֒:J7+7|vB׃?"bfWc+֩UWx7vL' 4j04+eʁaa9"j6sh [:\5+Į^#\y҈ʣGW\;QVfv\zDtџ1%z}{u0psVz?,?Wv' [ө &?9Z2Ȝbf{&x.?-Ⳅ!]ΥOij^&M>fc!T 䰂JJdV |Նt=Hu_Ň_zF:jVPl} _OqZ둖uzG˸ˏ.k}T-V S+p@9DL:ELox@NotV.+QD ɋwHZUEa;&Z mqSJS¸{z_w|<_kkwxw[ \~墷7谗-Gb(-yWx K95I!# D1ӵ*56/!Yykm%nUQ`?sj~s\X[ϳ$bN[hU>x_GuT(ei_o7ʶ jl>0?oul㎴ɭwG䏚A2'.m?qʝlx޾&=Css߿4*|B;WW.rJof+GS}Ƀ'Y\V ȁe ܯENC,҈(M^D&TFCչO'7э %e3r8<]Ca`k_zmVY;3o ~UPP1QSR-i؁D5LSpߍ$#w{+<"~t $a$aJu0K_YmOiSB|F YADC$ZIU߻zG4z;ڷWj썽{3ͧܝJlפj2R NXRYmWTiep_q]\/0_?Q)5FT9Jހ >(IyW5Tu /!VV)f|`u 1U*KsKpfN:1b^d=Sm65DS%L"yh5'[DMe=YX6ILfd~rێ/e㝷T!IF#F[ xrLɠ%6(jB.` .8c?d^S"F틮>?^Jnݐ%ٞ613A;[i> s'G{ Î˹7Ѵ/DCZ)E^ww>~uxPIbK_x5gE}620Ȟ"Ctq "D^@$i߳i6;\DB{b;11-1-~, 1j}D۲ behɰ"׎CQͶ6N|ѓA|e$]Bξ8z)[>%9eN*BќK@PS/=vw 2~rcxŋs7<Ȍnm]'vt v3r@ڱʌoTm ^WJ1b?'K-e@kYslfeSNg8K )RkE_ *P$Yapİ!ӕ |(YV&WTb5L**KUhG[c%ޝ_0;dzHN.DkO CⷲOvM>&y(i-cr)KQf{=(^TSFYGr0|a q?"U|ZDQ z'B"4*&9DH84i"*VtӨ\ٻ6$W2v0nlw]UmL I9q%L %QOL]]e\֑bW8܏uMW>iM(,L2"+uǙ|bB8oX5Zw*g2f95Vx4Pus ğ:3.Λs\ggP\qŝL %xѸ6HV,lKJVK%+Hv1pdxbWP<49 aWQaw50_лv;,O92;;S=`' kyԓӘ&h^!DNnLvV*\Ey*X!8"Xt!@ȔH[%B8J 2I bt"mth؜.X=Uԣf٪-Rǯ՝\qxز Sц,^>[3YD}r^ẀAKo mk"RjJT>|:1V+gbgi !hU ۙ9u.5Lr>gޱ)p[|͋>ѬףuZ752tK]HF]T^ULMG]$e%E1Z`5 L~<+/]2?|~~Ye\~UBZ.8_/!n_fݣVo/X茖^E# +U`8g|=8}&^VOlY߼h0دjnw^<1XX7eށQYK\n/:=ȾGk"FǮGgNt/mYz0DʅODe_ɠISG&)BG#N)1 t9nUvx!ͻL5e'L"LYˆ«j׏!?$zP+ |(Pwi%$P'%53C.1f,$c4i ({tDZ'i㌹Ggl\-\?93y画ɛݭWnf5">۬nv<2Cedalp3]jZ&r$$f(=t<{ ==|vGi% X|Bd#@J"+̄r m0ͭOO6#dzzvS2ٲ;u)IgE  DIA9d1\;8Mjg?  3HQn Fvė>nO4 g"Ԍ.-TFb8Lrl`PXb)᳄}0~HξNBj`ٮl>!7(gU1l3\gy RIמ }4Hod4U d92f׶P7i|E$o;?^jӞu8B[Ū?o8O~rvoʐC_x4\߿V,`htӨuU z_e5P))D-p}`BVm ߈ jI)/҈rE HgσWcW*I6Ix4Jc~?&Kӟ<=Mo%ͫ1'&2pmMpX *|FPȟ:0?@SUdD˗  g|`|>I_*ˣK;(ٙeȓ9sdɘ3?E;.oӶe4d1ɨGZ)դ2MRTҿ/;Bȳ~OBhZSAmm;4$i=oI)&Ne=v6HmQv4KM&ӿߗrBS/hWgʕezo\+86<N]Ife6o^s']wl~wՅG6iUck~ Z]{In_sU(͗$,+T k}:McF-BζvKqcI* M!3(-D+2t3d%\J:].ە9 oq3_ܜ<[c^q)&؋/}.HkMZoeMէRJ}WWe/dzJZ2LZA0am`*~ /ڿ_/~&,*|n20 nFdҏ~?f;d*2kE 2'4[V҆/U?kG\ْjE@ԅQb̔Zswyқq>˗|ZZGr(<ذzwCo^-?[-Cw| g|};UVW7( {7V_po|\ήcg˳Nj?NI:׷bt _m9,CU+<Hh IY6H|sLbR1+J+D!9Y-- g6EtKQ"&#}- 1((HfnU2PɎerg뢱hS212VKOO#v+o 컛WHqRhO Ej D*>!X(F@vJ.ս&{bW3:2;3,}3sUc {xJr"AgHْJyB4,\$+ {`ڗ&|򁫺z_º+Hև 2Rz*(|HdBTɨZOGSvU1|~>FQ)2ÕD:9RTb|V9pGCa>vB>6#qo>o_)DemlW W/h3ǙLa]R\^?TDQ"6a1,'﨓ƞA|R+m EbV $} (6@(M"9{}'4д}nrYnAdsYYШB=kgF# Jf!acԡzؾb_xV@|f.ubUĎ@N8ՁZ{>3zCcZ.HsOHp%;!t)D,TˏZ5"xlrx+E$TXOS3&@&"%(1eY@: 0ϱDJɚ5cH:AYAG5!j&0K;#0(VEAz&ܽNY{vz{޸ &k3 RdѠ,Ap9Šy#Z Sgj۸_a헽 4xsU>d%MR)«͋D$9mHKc[NW,34<qzh$뛹$73\=\ ;հsNxhxF/3-j!>O( ȤҤ(}rN{g r./Y>}}ry xvN|sm38òzR e f_\ Bz 3\#r#¤l~_P}k{=;O@4֑ɓ5훏Z{ShNQwuLE/Bųk(pL4?+߭/|惧7afq^-tŠ[u{c/rӶi e%[:kF73McI)jVCD뭢/\-nWaUnouɱVFkR!HEGK%ULtBհQ8& :,_ۧ|{?>#=g?~5_VNFE {NhO46mn(5e7 ?]#qvׄvQvu@2嫯Nr~C^\gaJ iYguMMarV6XK该*guB0݀dI^|[7{$Aw'I5+ @gb! QRP+IFH^U&8<= y!N=>:RmkRE{]mDelrԬLdt`$uTYlNe}vkb'*s5yXs7aҝOt'HřɗCl~Hݵfzdz*Q#xP~]$Rgftۢr|{#?wަw|}ǂl|by۫P6Z&kg _=ٰ(kSr׷LTLsNם=L髃]:L}dnps4 (N͏;:m^[} SsfŰV摻08w]֓U\jM-V=qUU$H6T -u69Fm \[:tbR0BURWqb>8Grb"s~ۺsvݧ7a|pQ'WMl7.>ީSMkSm=}RG[LySd)UeEO`զin!N@4Z$r1&Jp %ύrEy-$8U/*M]vk*q{^c}lg_nx0N1Pq^,?iMNϢܴRnZ^=3fDS+7%\Aڠ衇+r l*Fk?'t%}6Jkp)9>?ay7wƺ_foNlLyDd]^a4L'L˚Wv6l; 3ko흗yHϗ˫EY/[Y6lCj/LP3n򟵫 g;{c \ruڶ]Y>;636Ο_{vr;쐴ֱc~iw7 3in6(B_gRGbvmlCfzlQA|TRscLsnSs96>Z4ȚG>#{L gHkMy NY<q]u!k8-Q8vl{`qsQqm΅u-kQx,9VoQmpe:X!F!c>KW .C*5^*;13dCk?B/Ss.7GE B^Jj[Ks!D-NVrҊJ":|jax8r}FPK}S ;EŸȑ4/S O tZu)jԍd0ɹ%.Qa 0fIQH!QS#PkCv:vDiLZ%|q@hƄ ]dD)e[UEDG {ø:3mq9*'5$ʉs(j]vUT bltp DWp5d4j,y9Ougn!. +&9VY4:dZ{JC(ZX@ %X_;wMr8W&AN=RƄse_S ۡ'-leURR\ `&4$: )N),QŹ(Q7HS_%e:|=cYd2*,)̴޴XpYͮՀ5TE`܀AS@h!ĐHh,a@6{ؙ\mѪPFt[stTX,∛& \"00 Lƾ( L tf9pu>% qae6ud&;9"g7P@S/V )n쨃^yRͲVߑ@HU3ƮٕHޫ01j{RPR1JuI^ fyuc@H99o(D0P(6 -`A8Ұ:4]k?VV\Ѥq lkFC;-fإ*ʌYd1&±y9,xbBE}oj9NOUzʳ@Ӧ(ʬd &rrSrdJVذ938$;Dž':.cЃVsF<$922.2o8dͭ`$qac2@^BB$>y/(nM#@܎m}G V'ES Y_~^n 06YKaVF`;>.Ƈ`aws~'KS 8X'S 0"hw@Svd<Ť<$Jx—9t ʦ^H e@ К&6ܖSe>,)%ر,xkaP([|@c²&$˽eg:|lL,pP "2)Y+c7~\Xd ә4HĤJL `T@Ao;+8\nCIaXJ 1, )dlM'JZb~ L[%!TPgLkd ̬$- B)ʝz!M22ߤtVe"A yhC%B0%/lmJj6<.8玞y_ t\۫a|Vk넙$+RAP*n[m#D3)-zXm=E*"bk۠TkM!%'BA5@'  iR{ұR9p78%<%2`r[S|'Ѝmf$ڛIYf S :h B@%fdiO(hRC"6 ] 0!+FqĵQvb%r )I bX0p B 8Fsc %d$r GY;:@)ErOAm@uq19M9sq&o -8֤Y 6Rg Hxe<%40lb\;7n,  &%~%03&*ek#K=ڱ7sJ.vW"N.EC$ MmwH\DZ'tOm䠻pWsFڀ@÷ePX#3pdm =`e-p^u Ĕnx6@~FpvCcɪ rq 8)>H2"@S?Եvk-'j*XՎoݧg$<|Q襜; gLvwk/ ɛb;ˍ)i ʝpt6!m_\gagN=>oOieA殜LX KI{*)V{Bcջ)Snq5L5g0̙Aj90BrpWz! k_ Pե z;8! W#ĕЂ+WP JU&w\J W#ĕq+lm1B] RWR qcEEW('Bʗn;jJy@{>t4pjuߔ+lMv%oŢjnzz}4\@쿣+}G03V܅0+XAFJ\'K4u,{Jރ(=JPpeprpr(Wg?! * rHgpb1U P:+T+T]We.)N3Àr+Wf?À*!\W K@;bAr=+&wjWRr+ǪSg V3j\΋Aj;'\ݷp7 *V @*w\J W#ĕ3mAbpr+WV3;PJ:O {ʕDWKO <W1ql:DdEW(WRpjMATj̱?]%s7#TqI]o[?;j;_te]slelVﺾ7K?b~=[yjlz:_ kHSlzc}I iULjSJjպŠJJ-2`/W(e1Vq*&\W DABYtj4p5B\ylQ x9A+gF2yfThߓfwzͦ83&X3oG4P?ot5LY7Pҥq,>U9+,/W Wbpj=WR1O!ʉpbpj5W0q%6+lϝ(jW q+lˉP :} UjE!42;WHq9fT:ŵVKאr *(ؕiM)ւq[1*!LFHKVA0u jT4#0 5/W(RpjWRHb#\4ŀ+W W:;Pcĕ\(bpr+WVP%'\=\;V=2)g0gũaj噢a* W|8UϥR+,+W(WRpj-Wiq%߉b`ËʵDWq*J*nLIL+E)BB+T)jRJIURt*'wr=ץJ)<WRXC!yZs]/*&\ϤI?e [W(7WBhT\Y3ڞS@6^ZmE-65Sg9,jPղ5rm|.#~iE i4~R (W E:L(If2=y V++YjuPTۮqGkW֏T0,+W/z8P1i'DI(ؔ3όr-W6yfP)-#aΖ]`/d1BR+THJC^LQa͙O(Q\ SdR,]bUύe+|i=rnrmtNe؂p-gઓ[Ltjt@1JBt \`}ar=38L U:F!5^R2\Z)rTp5B\i'f@e3)gON&݁:+ 4 >bJTtF^Gi㸔 \`[Ό(u\Z T*N1:EI`l˙EAu+T)hLSԞ+>Lenv9Wpuߪ^W@լ\\cK,w\Jj)S@f\\Y P UQpEAB^++9+WVq* Lj+V+ݿ +9L>w*%\WUJfn"LٙvSag 7vm5r+\1g +U5\ɬPVIl*wˋ^| m _g+WA3Rnb+Zkӽ׫&@{}yX|uxÁQȍ$i|?=~lTfW| WKԊČ{(nW&Z단gAhJ}߉Pw{Ͽ=^&jYrcu* Mc4G[&ZRj7G#"o~ kp>[ꭚ :o׵fXc8 J+^Z(C䬍)9 0PLgew~kYFX\mb{g[h6;S{>>绯^-X~a~vzokoun>Q<@I#T.KM9 Ou ]ry|xT_aL]QZ-ClB;!H[8HZI :%.Jeql<X ں&pցs8aWεkRdN8sҋGnvT޻_u Qa^}L%/DRsͽº` k$kDj}}KV !F:۴5{}m7zl𸻹B2סSr/?/_Ct=Ƚݻ>4w _^^~>Ou<_B/\֋]n jNq6٧BX`Լ^~tRjǴηGAY&P?p \2ϼU\_1A~/]lWN8ogwY Tn:%ۧCߧmC ..Gk;u gEsP?r;ym5N|t`}9ljC%0hE=ҍ: &T3Mɿ?IK(\ a\2j*[BkfBK`a.)460)qƑ=NwgիgTW3hfL# 0F(L+VxGk},inD!.6]?o|+M Z-۠,Tu\ժŔMcjH=s=qXWiv̙̹;띲4-5̴aLmm2Px欨um 5EH,͟tM,B ʳlmn<^bko!s:Զ1#uv.VO7e}hCb\C!ZXt@G)R2 ~[ ?{F08`ko `wfY,A,9j9p}ݒ,9zE,9qb5E*qL?qV&eai^2?[i IkEs7Ahfoz7C밆ɨw'mSG$Oӽdr<k{uo>j=i\F(])'Mu؊p=>ug M(krv3[ 80_P$Zq"Q"SN"kBfZ\D~07EYp8Qf>:;ېwwmC2%ogMk,ы瓫ŏ**XZɠ>;&goM {(hozK}lS>yfr9"8[!^-0 rs? $jßn!^[AȚT{seKBnj]3l}3m^,*k!G̣Wpp7UsVꬓZ]W3Voq6Gr珣qzSJ|Nh/r˕PBk[Y>%~߿/~xo)?U8qa8[+C1O㭬>Pa K7ϥq8Jeul*Ēں{ o|WmJnևhͻw}EGôۦ}[*Sic,ѿG>Kv}]||ׯu^oبnTe6kO@u8ϗc5 >Zg Kõ2CU_cL}j2C_dn7?7 pqs/Om|S Qގ pw<2v*ᄨ\T5a Fx͉sw89nN)$uv fW27˶JdOƥ II:9N[b^PYR9kN9jᒠFB@l-(6Tc\X(.R3HR^q*ybK*%g}8Agί9渚STr=员0'; 3Yk p|/ s@_Xh|6>]l;u$%8@KyA%5⍯hK[HrYAd"P!Ȩy׬C[bס> ==;@dU'\ Kؤ,B D!$Ġ%P+!(pSM-`0b:Bax qgQez1Ifcq*wo+2ZuqёБTyB U>(h P0!Q3P J]Bqy<ҪR&FAQƸ&xR‚Decqznl $I1 b ɣxs24FَBWK ޗdXW'F}1=Ѩ-|tHS#7NdCz-Ѩ ~ NHXt*xQyŢvW1SC.Y0ySSHA:`^(M:0Orp{EA{I8A D=.SlOkL P&~"gz>1Qr,/|B0'8$@Qep>%k 0b g@2>(`4/DC̆dҏ}-"-lig;D'US- TЄ@Ej<1Ti /%Di8JIZ ΃**yI397sZ8;ܴz:iɾv]܈ Z%{[BuWTEdШdI1.W6٠R#\.. N,CQL<2z5rn8yN#y?Xh^qaX-!_2冯cW.oð#dނpF+T䦒LK1@Yi;BrrrI8B%KyGHk KT!'apnZ*=\{O7`\NAp"e:8%X/\}j kyZBB+I05u6q%S9 i4Gu}N# 8eYW 6tLMJ&Cg!9S2#jGs֗CWLx7uv/`2ٱ)Sr-sVWyTW}7̈́uIw-w=w5h'AJT"򢚌%"hbIꝫh)* +;e+mu)CsdT?:E$"nH@Ϻ8p;xf!H,:҄By4$+[] rS1"ިJX*gHŸN|PPBA.)1DEE@РbtDB26z-|&:.x7Ϭ 2U$C+tJ1L0K;^ըדNb8ޗng|OHQGW/91\'LS%bC /98>N܅\߇_~qruXi9%+΋F2E8J[AvmN5?K1$~ 7m7f7? .7c)|jN.e[v|Xsy77a~ޯۏo^[_Y5{k=;n l{ oQ/iaهg=۹7/(hi_movǿ[4 .ڋv{&0h%E.HȲ5&lYfxHէol)K!T"9^y\(a^%+cg돍8LHD305zނ f zuʦҸ}PAig7pxqŷf~/D\¥'2E0=SFܼ (vJ j“`PyVJ6WrK:$pGv>:i<<sˋ5'ui8 %K-L~8@F["(?!1J@[Rye΋>ݸbg*@r3BVw8W^gx['p^':~ĕncM(ui=7P[׏&!q[/婛W|W3'$@PL%uD3S4@3\Ҷ'³m1 =:$F*rh6jH3Dq*jg-qYhD'缣IP1ۜ%s>M|r9£Lt 0.ՃB~repukYpӚmĔ:PH ߬-G7` pkpa? TM"dd Ø`GB\6|9`+Eh4tx4ZW\/`lqܿq!=G|!50PM@oQ`5 $($I#0 +2lscćA?> #%MC=r7nA., ~Yoj46_?c%un&ߦyU磸rfql:h_k[[p ,kB+ov6?0/~KΛ]>2T1[o%8F|4ReϘLrTv8P#F~:?z6Dd:S²nY}u!7t9 1i9efZ(WO>4JLQ5?cq- +_ŭ]`ܳ {ž2_72ޅ4g&As'5\m矿]8~x+O P1Lst:s>oj޳N3ыݕi}߉6uǪyӵGQޕCkN|9]l>t%5Tgql.i*osEt+31e$#,(_t,"M`j[0=WLaN5Y 1d9&)!(<. 3AEHYW;mybC9m% ;Ϫ胛~%VB i>Jb9ROTUƇXm$BqR1b%DYU.쌒&/~EЌ_=b佴~Q_L&lr6 7J"Lx&O*R.grb*{Xfl#aHedUԅQ8r1 d$Eϸ>\~˫UAG^v}C/X~y]pZNmf϶Wr6RFR2ն)Jo|)[ k]@\\Q P]ԬիxdC_qJȸj'⪝Znv*D \W=eR@0\\AI)BtWRW'+S \`H1B\+T+uq*[{\8J  b+T+Uq*quߟw``p\Z L:A\IJ1Э[hrn 5\CVFhfCɇ?~~f0oDxc(;" ^[UEdD|ylsMI}`|ɥJ*S0^P ZY 6 JCؔ+l;-W(Rpj:Pqu2r- +Ku)B *Mĕ+S@]\NK@z=,zyFX*J0?P{;U;m pNw%[Jڷ$̲p_ W(K%۩UBF:A\1ɷbpr)WV UnEt:\q< VD+m)'^+Pi quRI (W(WRpjMq%pRq%)w2.d <p;䛛h\g>XIϛ 6cU30ST̯$}L}vWuQXT'Շ=%v+yrV:+P뷪WFz!rTxb e,dXBl)\ \}5-* ^)'rU9X|?yS+:-Uϴ+>۟QDʤ+2,(m 5,\ R %Zg$3ef3mrMͤJ8*b| c`B','=dwUM!ۨ% W%'Z aN&˄O+d-B>D!lJY$tJTJb+/BN = H)C Vt>,U~$ȂpiA(WRpj屶mW'+4մ \`E1B<ɮ UvmӸW/+U%M+gbQw*uFqeW k^ P)Wv~.zdѫ3bߒ`yZ=rgZ}Z؂yWվEOD+)-W(Rpj:@ jmLA/yq(\\Qwj:P=NWpBEABR+5 Jم+{کUB:A\I U>N( 5(QVȷZ0-/ Q Q-]4 bZYjyI`lʙU=ެZa+T1SĕBp +W(\ U޻:E\+iFr:(EOxP|gU>J?~ǛLE;U;HکYt}&;pp+bprN-)K6oVu-{56>,;EFvcQi8dz?G~4|0B(į[dt𞚿ko-RӧaV}gƿ..jmrLlIUG'id*o'깢Aɬ*P|d;-F,I >C6y-q?x(|UA6j4ZpeG!fH&{ &:YԆ+ᕗ'U\{D- j'zMR8'.hۘdR;# !a"k eGͮM%II{?ӠZdBBL"u6֧i[I\&jOH1@vPŒPTSr)EPRR>( KHA6ՙYJ{vGk\50K˖r2DϸYJe@2܃'%V&*\"$[ uLaKIJBeb̌) >fAYx6s *$w( ` #4j0hn=#H5]6d\cq&:sz0jc!AACŜfVg%gĤa V[V OSǙi{B?lBebq 94E5zrIe67nHAhN kU$bH$8I֐@.[#4- ᡧu_2j!y,*YJvu guuh45cPɁ]j-t-%VR,Q>ՠI @HB)BlϹL*&&kۍȮS̡~!` @6!)[~gE=4#O%b\UkjՉ^*QTs,tR/.E#OAQ ׄOQdO])i0f;Dp+ ]+HW&Qw_KͲ#oG0L3R/3 E Uԇk\4`%[ؙzUPQAm> h-͡]KvXqc~0P&ܪCZ]/ .Zn kBhcnu+XLĝd] Hqk fCFƺzzEK05n0 Xa=Ҳ1 k9 g`UhU%`}5 eo LIr,>_lG?5mB7BJJJ#L2b!]AАA6p.F%G!HTP&׀T߄L 𯒡2UWHPcYTB29V^X szeY r3.Tk@ e z0 Da(!$( ""*f"HMwآDy>UFݚ 84鰐A ;H+fD$' GcE( /} Ui:/LM{;Ѳ\|P]Z01ս.0AABZI|txḰKy@xc#TYiU2Mt%C RbtLECNZ, %tF\8gP4I"i Yk2!`(`-P&>yt_V 5ȤuՅ:@vdm^z -ʰ~d}%TU;+Q;-,F ݚT$I"e56v( =sx4l}EeFI*väDy آsE6G=Tʍڪ{,-hr%5 2TA2 5@J'2(xEշl֣bb V$⤩MkurSp, wH._* Q 2P(EeQAR܌EEH,{ag=uK[sOҕȪTqQc@sjo6i]0s0@ZXf=ؤ=Cɗ U=e2 b2 D hs@:'/y~5kW!*Z>@Mf ScUwUA`NZS0h¦f@ 2) M F:`y7tm͢BSp*qK:kCWQ[1pQix.SS.\-H1ˡ옅jIMH28DԅK/q$@Br6B]шw9ЏL0H5zAoYܼ=5TVNS(Ov;$ԻɂzK*W?}^NMw7/ԓ)ꠋ. ~W/Wt6VRlKy|swHK:K?n7M:.ja 0 z7on\|)?##DKcw7_IqMtsyq?e7רּ/n.mwn6? q֫ ?;I?:zX&e*F]]Lqt'x5qюKʱqKz >r"aZ)PJtutDW8aF3 ])\=]Jtut]^DWlT 7 ]/ NW@L`:A2^>_=Cu]VPFgN$mg|w@[xuF̞o ܶnCMl=\Ϳ _ev{s֊޼|9QsΔ&z:?8;hA-EWQhv&{sithy&^BW62 S+C#a8 ]j;RZ+@*qAr-BVOW@`:Aȡ`+b abhY(Pz'}<<:] Yp8<_/jʵ3*2]}(n"!BW@dX;]J噮NTT^qav"JG , S mvt1*+pr {8J? ]ZNW)eywsq;/vþNW5ob͖.[B wW u{&/oŇb釫E~yZ5WdVxz>n޾/.p~zw~hO-6J^/yZ AT^o/}S@B/M<8P_ְc>v3QƸ=P|O>׿^bw/oૻ `wD'e>oz+KSWMMF:EoTBPɶH{R:oViyU ˓B:e;hy^Y䈇09ӄC:pyGMUU*XùW}F|eۛMM>"m.sԨ[x(GmxlF0t㻏]LZ<S=8Z1 i(FFZ %~P(K @>nK ;NpiC U5+Bi5 Ut@tE ?amyg(\~ UCWW s,v"V0]9J )CO;KWK!Qpz,f6Z}CdzQ5A>t9C/[5 ]\'F+BPtutǶtQW7 CW@k\ sѕt9@t8 ]Zz")ҕQB8=]s\H]\3"֮eLW'HWzoe}nQ\si΂ӃYa^f$QI} QoaV~+҄rUw\2K?K;evn"~0,*OBDWx(tEhr:v.} tu:tTq8qaК*j: QaqRҸ( ՟'<p|Xc/σkv5;]CWV t%>w؏B بaZ7 ]ZNW20]"]) DW8aa hk+B4 jAq[ gׅQІի+BY]"]ì]vtE(c:Aۼ1,z21{~U?efÌfeOy|cc܆?{d]Ix8yAy'U nouuT6ҝKn͙AF>@n1kwO۟ACrepyE۠b>(RwSjZ,u=n{ĵ'QO=!wzՋ}h;+Dazv{y/7?n~{?ǻ39m{ǰ]c_{_})s2USO$7/(eՓ]c>=)JGs?ǿÔ_)7ԪfU-;Ξ!jJyUW^z "CWQv"OP=ywox{RtEpV՟wJ UFK1]`3k9@hLWHW:oF:N8 ruEhʵ= t>qՙQ] njc?hZe\R3J1]}K'kmIX~8$&Yvp;;;s҈籘=$%ѶL䘷8l6U_] m \%q \%i+qRr> +"(bGpV7pմ/puJRb5 +*8CGpq7*I+H*I)W/T̞ĕ/pմpRr<]DJ*;!hL3!Z2s# #ɳ' Lt%4J!" K'8u_*IKy*IٵzRPݧQзp/pU_b)yR>%e*LW -FpdWZQLVYbq }$c<rXb~޳R ̎7;N\puT<3xcS\TH8JRJ>هUk\= \$=$⽁$ \R'eעz"zWI`g0gńt \bcݧ+Jް+"uJRb1 2%qj8]uTTcuǶ_ !c"Ƙ1)5{q8ң={Tȇh}̫tƴUga=i3C|՛lG,n(=lG (I1ʱj>%9t~]}"wU@s|Yu.@'L2, .pw]^?ᅢߵ4 I啫/#! kp_ӌYը8gc(C[ռKi. e=xvABoR]̗Zzv3ZQ?PNHYᮛ7-r ᯦{ޖ-ڠZlHpy=NXݺ,o>M)-]eӟ׋wkK>t|ۻ9nNY)ZѪ;u5!krI6*P[)j*0u8PN+m$pa<\xe$uP p\M2gn]3 ^\Ř *`DBXYpR#RSFDIH0<)c"7iIz>T`zhe0}~rTT #vyt( ~7^9PͅXo׳ekzwS.ӅWfy&iqpDǺi7ڟƯg%"jQLߛӟNo@Q>yK9qnwtO2('ܵ#Z}F^ ) l$`HD(IV!s1ZWmՈdZ=!"ă"p0$J7lbi1g;Ja8Ҳ&a#^U fu߁1'hpJ`;Qhمz{%p-pS`EM/պ ^u8xT#&=u/6\eݙ?б-^۷:Oq }y'^89 ʹr'G3=~bl  b60*y< ].EIuoT˪Tt\ @]LXKw[rgW?aT`k*=_gyiZMCвak²rT3&&{1ÿ=wB6iw}$n;ݜr-t/fڐu*U^:Go*] &0m|1˷|ՔS>s4]q C!fH rVk/;^sp>cI$wQaq9͊| Ja*g$kݛ/Bcԗnh:KG.B>ZB4B~i7Kgo?OyT1m g[xl9;DvSdbLvWc~GsbAiTqC ՠgUGS[ђSk((3Q*8F,g| l'ÜQf YcI"!@Bo?EpڟH{D#R煐HB`H"Q$s`)#ʹ4yfNQ(7=[֑aY\ -J)OЈ'ZmXGM͜/m ե;Yw3 58A2K&}3QBW׏"gL!2 ,%& ~ A!AH{jĞI*# , I`P'J$+9 .Fa Ky$S &.V=q7K7/WtK|j=T:iDО[Fp J5G*IVsT ix[1ޚ}tuk "OWsޘo-'8/;pf6_?FT./a̳4)I3D1t 3MDP*13*M f[n tآ:-Ybt[OμCf 5$ռK>\Sާs@`*s $+9E Ĩ%)pKwZ[D3`,lF `sHr()G[ȅ4B 49{Ҙ5qaԯBs0v[ӏSx@znX`8 zL BP,3ـC*-$My Q+ |:›Ą3[XA`IrcI-v}G!CBE&P0}#l/ɀBwY1s;q KPK52Q s`gP>99w?O MknfwhûWR$HqBP0o!7`(ћr4Y͚:Un/{'/R^02t4!)TŒ~*{(Qɘ|s?."1x4Y#0}9ШE@\ u/L <ޟα*h,l>fZ?\ x RA޲ Q`JO0X]PHX]lG`XFdغ֧+j|X=߄~GҴ * n8[&c|\NF2w6SCcF}_w__ͼAg`g܏ƵQ׸Eh`Yja:FPȷQW{@U^ꝛ^~6T@Y CNI0~7ghpG =s ;[=t?[0TBkrf_rfp%IV_ 'btFuO"%g7+Uз/0gW0Opz3gȬ]fjWQpP,/ytFt]Eyv @u}`ú2(gu(j_ar˞\eGN( G> 03sL3MVŧ<ZT{E* u\Fdh:z*#D-!2E~aieXD"ZZkaiOOE8`Ep޵6mdٿҗ$5U$(P$-$5Ϲ}ځ'38bA #dNJN=l<Z~A/@q\_*Ʀ jOvŅ{xIF'u+w .=x2-7Iu&_"hʼnfs˵"7AFWކnvA+Tvc/WI=hTaQ)gXǔ{'<)ZSlWGC*eaYEĖOzf9A ۖj˜u  TO.j6^~n`_t;Zrк`f ë;«>^;nx^TKy (60$I׾n}tnv 6q5p3&& V9"D[6LU$Kz_\h Oqln@ !k}{#*Ls14Y8JHH 9.GKTͥ&Lm)JFYk @%Bkۖ{m͜5eZCSS Fy1L0-'Fi1f8p :VI؄F(ڄPBz* p* XEd"B6xIM OXJ/+7b`TܽO>_MH`Nc5ګ MF3Tp3Y+eXc{cR!ɾ"o^poJB~Mh+%$tҁIҁ檗rupז>"ɹF:gܪs V ]=c|""v0EZd^,Up%Oۘ:Ah Y+j'4Ѵ}σVV7;ktEۄ\;y(Gf ID)U Qs6mi.ґ *ecs2JecJ8dG$ =(A4kg]+kax( @dJNE.4)I 711}7w2߾9Da9'@X`\&QCg]GHyt"(O!`  #`+P$mR%a^B8_ mblhӍH(+}b |P:sF!QH}OH%g_~8ITy~>S\"'kS'0~<ا X/6jcd %eSDbx.mx< xy:U`dE-6@5Y=l. ?55BYQ98?K( Qŧi$靤ٔ Mz9fW$U/AϾVݯD''"PIn %K:&?hJɧ6 _aHÏfP9L~:+/O5k<8zrQ8g K7EٷTv.ACmu 3vn? ^3Zt inFfˇ"F KA4ۺdz6{ɒH5JVljY%DO[PHji8rIl ~q6,+űtSETXެ7(Ogwu߽M~xsw3L_N>{w`b8n4A=لߞdO0^}#m5 MSV7Mj_]#70>m =7[ S7co\~_=ǖ٬Y |T"[N_:\uªZŴY.}B`̀@kZf4j5fM<'.$+uoF;BzF Wd#jLu<u >\/0/%Q') \j ETy-HzbX2LI1taxmwkY`D!e!x0ά՘1jye4zl5BZ"Zƈ&)c>m`2-r T;6EjRl\ʅkpr]LbL4,W➕F]R?KM+)JW&*i0b29Ze:]W1ucoS^ٷ]ֽ< sW.p곫5ɫqƘzIU7_qcoeYUIHoy^KM6 9+޾JNn W!1yqqs-(s""Wqs3jsWc>gX"(#v}k]9wW{qw=z29$h#3TԘk;J`rwˆ^ G"`"RCD-a$EV 1閹{kGָj4z-w@0s/Bx.[>C0\O.k }d=6.- UJ1LBkw G#bLynڕ݂86@"h@Q 2)oQ9b0XJBԀSEL۲[mڃY߀' tֿ^6s\@(d3官fev,Y̳b&z) z[pÑOos6Vl Qǐ.y}3~w>}~NWQxD4 m0Elsyhu"k,B-;y-WB)K|\'P>O3{Eg4fr={ d\vYK!epY8јG=G?ziL8zZˌ\UӉd"7hr+8ϩ$1Rq,vJ3L4ӆҌ pG Lc.)Xc =8 \DA)՗`x3h@Di zt}PkzS8Q_ѺvKRܗ~咔9VD[n8vHkGH(֐t9W.|@[X3sT8Sx6\y,qXbaڭٱ9Gw>6KVr..ʙm`huء#):ib<Q2\q 1sY`2CTH[G‰uXVN:tT;t(\(wp)BcF@1gz4‚Sw72Qfo Ś礅lOdm}-_%k- Q#H} p\Vnx `-WzJ!D4 W`.&\%sdbyJ &f}[l ~d8:. RR ֒dwQAux$+q~z p5y} %9RTE0N?aDsT_<86dqsv!L>^n펳k-S:Z~7l}5;F+9X{ۦ?+r`QBwg%%#ڦ9pmӟ+UI {+F]z`(:p M&\C W`vb*Y)e^`XHtH*d.rUa W kdZwpv Ը]p1Gsڳ(Oⷊj(:tl(Y+7 w-m$GIpFg;kv7ncxHɺUσCRP"j=U_UWU{U}W`RxdUW/Q]I0=&tdz!} UV<*b+EU"XQW\y4ZM]] V]@u 7̳1eЌƉ-fNIrL)Au|v?'◽@F`cfppBȔ1 'ʩ|8wt -4.57k.|ݏ4 Ņ:_U3>{} '¹OVFrETy-!"@F3C+AV|%^]@o67xOw}`0MBw6Z (foeǿl:yg&.)kaq{a(Z^Θ(Sam yapD.L1P@8q č4zCQ{0rHRIpxaR[ERcib h͕\o2uz0{tGRf ΃6!,z"s*C3@IK1NIĝXZf|پoHwD(%1 P~̰\6N+̣A$ x=T (1Oi⹅u+ VkDWiN&焎AWB#Xjp##<ѷ,ߺǽp͑  9i (XGm1Z 6c+H Ra!ua/SPXDH) E ڑR Ѧ-uX1$K'W{33g) pN=_$;~:(yRmb$?Pи{wQ9P}.}Ox%yܗ,p>Nc 9eSDNRbNxmt<%x}z:+U*aYi JwrZM. ?5#tEY"qQͱ<1~=Tg_]ghrER귲\Mg2H)٤7>u[=]VVFSX+` KWP;Kc*'~-|oN o b4s6_^^sK|n$jбW?^f$2+{b|uOuݐn+YdO >LVAwWMxg;Y32~ed]uce=uT 9_#ANS}RiTOh/nw\wx~tOo.oD]Ϸ>U0$Iu'4݁oOZ65WuM5EQM!B|[&\(iM߿뎋=ɷa/iej蘾 zVeȕMT>UL9U'xx;6?bȢFZF$`I6V82~ HJˍv+@ (G(*jv8T*/˥!` c(j<m",J˄`qcJ^)}Fua2Ⱦ%?O,ЖYL t_c3,gV;\($T<]uM7P{#Fg+wNE|ڀ-#@ S8#ٞ, 蓇Km{h;UgWR;/\UTIu3'Q?aӱ ~ݴk FQ0H6D!e!x0`Z1cE#^ˈiDk4fD4#gKXxo 7K`o ׋T;Bu~w`xiA# hSXy5xt6jOFW%fW*HWˑBr#XY '8ӜzY*s?cM%_܉㨣;Bq)8!3g/&<5B P T"j83ϛjA'ue&EqfE"㢜şnɦoT69݊(c4_~Fট-TO,N.W/j]$ x:YvrBG)]upSkw=%`\iխ.ߢgXn< IL7Mk`˦dvJ{[$!7IS_po ʕ'2x:񧧳t=!kf jg8f,ƞpIXa+$JAi6#p2MWXxp.`KҀmQ\ u0$QV)Xa)k ^*m4U aU8Rn+帓0tQc6tzU i v:o$Z޽}? D2aSt{W&y,g5c{ 7'zgk߭fXhsUضHkCZDpfQ/d\ټv痓ڒ"`lPi;qx(ڢǢQh}НGdlC32QX@#Mr9ci!5B>x6\y,qXba ?yfr6+C|0OSхK?n7 y Ћ(Fm_>ۉZ~%Qч EN@(K@rwKp`V'J0r \H Ks8a,JBGX ٴ|po Uňٸ;?_fhKozWTtWU!۟?vXiqNQ":ϩ>7E.*RM7t/OoO{3`A/T\b2-_q]/ⲉv38 ,;@:,Nc&L~_.o &&VQRk3@cĜE NV?֪u?q-}>z˱~MZzgIg^Nt?[(r -V2}&˔پJVߦǔj_ `:kc)A2`V(+td_po>SV"EXν#,hF3%i& V0}{FlFF|쵲2aDD&p TQ͍QmQ'($GAآ%[[h[L1E]2q.`9MF^HKY$ |%g`({'ְ>N:<^ui#[FA_vޑp,-z&HRON3Ӄy 0E]YoI+f.)K@1;@`ڽcyJEEI7XD%R*b22/"`Z,q$Nztj%˅[> |>nb)ZsTV#퀤A Idl |&:L{I+ClՅ 1$ˮ& xZĜS)eYoB½a]Z™9zD>3^kM>/D} ;> IGuD抲3[HTs ^6MqHHMq٠u#'IXMJIIy˨ĥr@CΰjR 3),8,YַMZg= BL⫈_+*1B#ÆWe|D"սw|5&N=畕TU6AӊZ94U ')Au۰W{8~x/>Ƨjn[Fm߼~C M>Z1JLШpQȆBϴ>gCﴘU{^N];G^THn!ŔQƸ"HQ +7"tq _Ilu^*r83|I,㴞R䣣/D1bʚR6LY7aUq澌Y S'cw^: B$u^\WfoQ/&u*)P鸉^)Y^R rYHD&NeD;MsB,kP&gMN$MPZa+!{h(۰V*۔AKfeUl2FIt{Mql`B[  #F:*q$HLBS"Q 5jjNѾ>fS{ɹ%**?\ fT4g;J H:;{D| /5@тkV)YԤ!T^ q,tJ`Q"I% AmdlMWi [b!#-cpX8(mrd{}[ ׃~5y初Mו+K`8=WNidPehKeXP1[{M{jw vYKc>D^"qD օ$TGu`xؚ8Vgni[cSD-#"Cĕ[N}<9Nx˩H'i#1[n4%D*@8 $&q5D%UA2Z'ԓHQۚ]߭ Bė񼀛77j/8GgZ Cl$h|Yj&$@:x70Ihkt+ >: `bP"Q ӂ .ݓo7d$s-zN 1^y~j&gߛ-|}vd9 P;!X;p->Q;r嵦^P 7N佌7OѻL6Igߟ=.Oj?j2{+#>Dz-Q޺N_5\?wy|k?-GyZSWu n8SC㨖~ytҸ}̋䧡 [bgz^|Cu߆cG ]=ώ/e8eo Nv?G,k|U\_p 7Y[4u?N8p4 }08_VƧ D/iNNٹOCe2;'I/nd.f4N~8K6#\H7]{W-A)9bT#*O !pU'-s1/-x M{{|Vkbd0%\&ULc[4>7kFwAoVWB =zD+^L4|"4 Us1dmXYi3l ֧w,RQic jZ?d7τ-ަL4^Yuo+h7_X7ӮeABa,3ZfP 3[XH$w+˽AWQKeRP:cJ*.lY\Lq$tV ՞yұX GoxjfAxAXwNPB+%x6Ak<}:+8KBƦŬj=VpnI]Jܠ"Hd>8(-ւPD(IQ%.z7AЌtAvUdEnA\L\\S*^w9A(2P.3Yg"wrNdWv)I@ugCx:. `DK VH v%ck\~ϣU[m#:}K/^2A?7"_ dmV&05iEb:2|qc'c@(ЃI42%JwmDc(%P%{8p P`.  \eiwB)Žہ+E8 B)WY\N;\e) •@Jߤqy9W2׹~k\e8aQް2f|N"tW6=`3Cߞjϻϣ(8)VԼxg*n>BHࡘG]@>SpUA uvVhap!ϞбųDNw6;1[33p9ɛ-L+EM!׿wiHdd +tRny*@ݰ.S)ً$]>,k;O/V0Q/__?~x7m5ѿ -Q}z„X(&i4~1IP&s FƆTd,gH϶(t#ۑ]a%wH,"lq  ifE4 jb 1(A KH!)p]b̩؜HBȴ`L Y\ŴrM ("3-AӂK}@pSWY\WYZp\ER`r@p9v W(Rq+-IUXN,:Қ?gF)@Wo 0Aó<kbf{#;ěrX7QaN)٠iݲF>!N=8?NWb}CW;yKzaX:bt4` goUV f2;c(srP>C@D p>-=iM*U-Av])xH\kJ=!&ji.*`wItOZ#U!2[Ԝ˷!sPyxNn2 {h<X&\v ~CLz?/FA(eC?_ˋrx].Xe(i4|4!pn|r$eQ`0n9"$v޷tb^&-(ӚB{uhJBIjm)+"@B/]mo#7+|6_o6}\`,rFHL&-cIDQ-vu7Sê&$"a JFrѳ}~'v^iTZ58D]Km`4y@Z;tLVɵ[RwVr'}PVKS}7ʌn{vLH3^b2Qs{)k(Z} XXg\C,%"|;{b꾨);֋=sK/a' 42Mw<0OdZI笑:/`|}~LZjEZZrY9~,sc[H03E)%Sޛ=dvZZCѽL-Y>J7cCo?cɂWZPRL8q\8-n}jM~`.Gs&`)pB) wܑqTIVYy$OMTOWpCd.B JP@!TOU3MF==SŒ!F'?޵&=|[}NK!EGy\ of;*ò *XK64V g',e1͜g($@F ݂Օ9u 7%q9; #DTu ZB'-:3L #KUNA,pCK톟d4 wG:(uԢl5Iz,JguO!%432ӿ*̂57jrgwW ;U&tջ}x ;1C[zL?1WsbN?pbN0`{WeNyxzt ¤偢FX+,40t"^_fT?F@mdE(iƛ"DYbYcv>EKAFqTHGIWD euT,%EF QT3g;BǴp@c-mq|"󚝅3z/nq}?gч>By0RNy7m@RV e RFs& }JNaOlϛ HYČA]`.x (v\6FGοrY&eK6qH ()g \A(E}M&Nˌ뗐aQ4L/0y7n>o<5U]Md@^PLGUGX TheLUco_Hn)[Cz0xfŮ.}|LLdz=flLjLH)AuQXmcQuQaq{ Rt'v Np->"spp2&4"2I+rW&nVآݝ-Y fMIog/h* |RSGO1b Lս8g=~KN1k^SQu켋A4)8 SØ!jL3 ufu#ڷ$QNÉ" VDA OKT[ļHh*4)D\6 lh/2`!Pt ̰hй$s&-fe!hqP}o3/?#NJjṥM2euRجrSCa#vZ eǺЁ'ӁkbRْY fcl-%cnp,jd'U*+_սG;~ AI;ou6,edwlKK%nuϏXUT޾Z (p&9e\)cd|5("RYHL(tYsL<ȨɈ9Vrjk1 R*>HFjlGz\V;b`XxP,\XT05+͆UOhh4y?ήpΆN*> &dF|18}ICF!.)Raez %( AHQ`Sj5 hC)ce.9̢:jdK!池vq$ݵףv`u.D >$"\yσ&6W2e) T$TU`V"-7V3g;A֭Ԯ?vED^y="nb8Au~,KJ^r. ]d.X F'+gq3 PlH2]I  BIs#"iFz!;9L_$zՁpqtuLszɮ(+qōL{4:,L]se9E :Ae` !g:hSI!phxbޱ+Be<(Mx[#ec:g3W?> ZOj>y7n.#ҫ1$x B2].v;߼8^t"eYl`SRlpH*u#ZCrUW ͳ`]YIv p|J}4u{$,l[bWMνQ>/|NaKkP'S,o +v?ێP~yrׯon2=!}f2ڰǢ%Ns/kv>Vi%]_i&ɲr:4$~֎+,BhWgW]F_5dM!}ڝ+?w>L{ۺtlA4 Сݧ:|?`Ywv/`>Hgs= |пrG~@ޭ9R>=Po|L,sAGf=^F R:EMWC)MpI4ӫQ0F1&$ g4ZUr)U})]R^OG% 7Jv'rĸYd@ЧC,JE6<ҕUL.}똋)˻o44P4vi0~i< O]< VM9dݫ_'kLۉWY~gߏ*/.EZ&/jKZat%)ơ)qT;ro߼UN1ea (lTDӟ3eN=YG;h4=,zG f &B6 CNx `d^:>B٠}9$u+XmcJJ.[ 敜yџ6W}zo;ew{e^)tLASxɢ2!s}`F8QN:=A }2G B]7gK fL^Ok>lI9p% DkR u =܏5:Z*ލ͈Nj:.Ztf:=h]३WxFECWQLqIY h !ĒW$5@:3^_`=kG/%uGk@1m<" uv,_KёT1`e}]jW_Y{H h 8N9^דg3:']S'k(o11 C6^r{}54;g]k|q(dhQ"E#an X#ėjxqϛl/Xv)EY3хD|.xjY: X{Z=8+̑]E3KO߷Bwi4[soƥB"J?ə'a􋯊lIWGfV#ъd1}xC7"c`x+Gh_ .`Nl0$f[_gHw/Y [azruEq |s}RwZ*1l"N& N[# +(>>'W 0'SZ."ǁuM/\:I6x?c?&_ox>5jۓpg?ǵ{(>^u EK206~GS?ys0ϥjUӫ|F`|>:ڝEdLG __8dtɸg<]F\M'ד8+~WFXL:4?0{7Z1͉:1~ky]_IhZ1)`ep%C{w >^6MfJ߂FϕT⿯0Mo_r>ްe8gf+UwSZ+lnhL=A fF4&twE]uwktxswj:|סtwMTO[󀺳w- 9/#Iqa=R㜃85~JP$CR(1fq' j[=1)JRLT5Cs@C^5X I[rD7&U-ڬ^zpBc-II: <_ fMq;6qZq('$rM(?4 A?È=a佰4(X.U}iS0Q)gSuǔ{Oxi-Oͪ:=VQ4,1waؒug0 ^t!lp_q?ԪJ֝}^uC.V,M(au])3*12*ضL7_w4sZ&d 6o̍"y壛 RBOL2kߩOu ($4Kw nq4D=:}hxX2ЯM<\VȞ8xׅOѬ_׃(6 MO]nM䮅 ?Li{!O[mwbߧ x⸌+reϙe,A+qD{OfV˚۳a'"mM@ 9 sIО)F$f[ps\bGr:kh65d3}} $L)]M(G z9ճy\^:n 6EiJn+z,K ύ[eO7|EZ,H:h!rF΍4Hᥔ\͖]6]Zы]!f~A`H,$-a /^ ZÅ<LNA{~)MP04XZ*e;Q!(8Ne ,RLp7ܴ"*EP^jcMWQE$!A7R4v>7KxVl{*9u޺av0&$"*=ìRR%b$pn6>IL `L`R&1j`(X%(KȳJcptUAҞbp6ȓg9ny &8!0J31ÁcKt + ` =VMh|EBዼMjm8. åh2Xb#ENjRt?:څryUF>I 4ϱ ,9`*'@&P# Q[יs*HNp3Y+eX?Z4Ǧd mvIJA;܁Tah&亯 9<`ADstθU5.`LXc0ο=8jSB$"ȀK\FqɑR9G,ZCJNxz}fZ֞halRbQO`SE9"`h3PR1RjoSq8V9ZZٶ\oA:R`>lg1 s2JqZd L (O3@2A֞ܮLE%DFxɺ VkD!9c$1$HO|9D4rN8:,,/> #q)#bl~"5;*^:@Ub &FJW(Urh6la?iO8>)|~](?c)H  }b |\)uYQ'$s|?J)Mĸg0[TOj>)MTՉ(b#y,}fg+0S0EdN.Swbfpe7'9I꼺VW  ɊN[lP )^ j~\2pE\G|]Oo~ ȯ?Y@+8 8?|ah|ZߏEtA(u*EL $V2j1ϊEp "n.8W⛴i}m0µfhuz({Ű}pmf9P"],{n~QlR\̯Yot4~mV.Ī)K-3j$i,ӧP'AV __ޯw`Jp ]%OMUBtOWGHWg]+̹ ]\hW JNWR+:BW WJz:B˕Wp[={F炫D9vϲ+OD._ɡċviz:Fs/ 1F@9CkV+޿R"%],igh:U+4Юk)M/$z>BR%#gPWv.JUBIhOWGHWRx*f%3; l.T|tSBtuL5 ]Z&Z# ('+Ctp$3DrJ~9t%|i$kW;VwDwK.h1"/d reL[AWr=]N:DWXs$qg 2ºBWcvJ(W4֞֨N1]%vڧ>{p8]%JtutVvݡUB[]%BtutBv$3t2Jhn;]%׮0CjR]2qjڳy;0 l'0Pe:Gs< wJB0V$Wy9f▱U\:cXgx6颳E*C Ƙc|:N]n\5_1E:XR&D'pl`2OUZ(ӣAyLۯ߼10 f1grH8+:HzɊ{pg;HdQ6!3Bb;kADE>ג)ER ?Hb~j.UOa ro@~v?HL%!>QB6&)6gF\s6;˜<ͮov7[Nkc TUp6 E<\CFe'P0Vgt%sHp:YHg!՝qZd= %bѳ KA|X8B%UB+['SSҕ.m3`]\UE6\ xJ,L]2`JpUgVB}CWjWNVZ|T.źzvK^vCK_(o7ePtzcT+Iw*JQջ[kW|}=eUtOV:}0;ˊ[3wՍٿ^eӲ&ݿ̽+BNZ,I֢Y{5]kt63[~UUp8vfׯ֛\/aղ:75|6OW%JA?-n!We& 7Z'[R5gTV[ bLMɺVq}zoj­-ĝ'MWY"{VR *[\N,[4F@Ez1,a;a?,?/<4_JˬoEUL5Cץ:TTPJ ELPdT`'x`ᰡ .aZW\cWw mY}zs2 ҤUY#>p$8 YPJ$pJL)\gj㺑_ue pafa|٘ | QB}YϓDI.`N,<$`up=g~ꦖ( 6=HB*ٔn۬>ZmdP~ƎRgFz2?/4(ɡ☦# LƘLLj`|eJw78t]Dr"c kc16T1:FkoL hFqNe֒d_C6,E&Ιc]|j)j9  %Gtre3sƜDh1½ 1=xLf 6P˙-d f3\3#j)&hB0{v MМ5jІl;7,#lGt[,:1T`)MRm`1<>gm 5iL TUQ"O^7,'3>_/ͫƿ7٤J#< \,k!r ,$b!OE$Q4Nl/'H%Ph9Hnfj)'@Xc{K9ɇ-ݒDBe+idJ*HHI͔%15PX!`qtC*GɆ#D@ϦEx īƭqrǒ5ӄabɖSY&ǞRYc&Ht=KE Ȏ^d^i"og0pdMȗ EU%SrM.Vg;WH%*Clͮ9LHp} BHx&X(%BjOHPJoBe:wHPcy9lA2YK y%ٕ@r3.1t@e ¾o}052ePFdBEAy: o4)i+ZmA.ŜU' sǤ< !.A ε8bkY*)ԙw D l"3q_ [betP;7)tiU B!U{֙ΈRl eA!N]EA-Cu*R\9H]Qv J\C`WGȷF;$$UWBdBj(=%5dY!#*DbrRG4lM#FPz9S|1b "hʲAB#a^^KM2CIX@QMiD$7]y;3¾]/dՂ`r 3?zPw fLm3Ska&a@ /-gu:˪dI]t>SaZRUf&1,:I. 'e>Z96wթʃk IԆ"9;d^0P>DvZ7F2]^"h]%0A%- s& q'2Cp/},P -ّw g#xa;PM v:"}nﯯ..w.!Oa6J4`V>Yb-="DEXjm 9(P"Qr@݅ZsM 8O5Ҙ j`/Kf _!n[uڕ[1FcGX $A #;Pj^v:m)U5/H#V稁'ΣbNp&B"(k@Yc "tV% ^fd $<+AjDP \𶣆iURiAX4qV)ءhkj-ٛO?hz|, =i[52kE7MTzwƉIf{ImK`#ƁC{+n,|Y.ǴWsT&`>'r 6chb6s>H(bZHutkiVVPBʚ%B{Am1F~(g=<>iUgS5n(!/G}/vdsC ȘŹGyI,ܐz֡am"K{PJyA`=u^84Vo'"X$@*9Iup#,r;yUySla5è+UPX+eQC]Σb$77cY/p`\3c1ն0ٌ54'G?,10sqJM@zXf=\U;,,1˩ՆXM2Ԥg&T$ePbI[l$dipMl(* t+Ayz PsHTnlr^ߋ[nnnǍ U>v~NW*W/dž^ca[r~{7VRn]54nv+Qzmwv>.oo~}~?c}oI?ܼ>;:gl+zs=k|[o×[7oж~7s˛a)S&z.♥3GWD-9Kq>+}'ϻM=?vɂ](eblKe_-ڽu;˻2o뫛۷vwhOyG̸3[o~ vn M Z{E|ptr͆ ]ΚzBWӕ|P )]]Ŝ!6DWv;kv7Aк+A(Sl-=brvA6A~O+A'}~nLv9sz`u18!}[zգP(*?=%`b ] \gBWK?:] ʐN,J6DW ] /m=_~:] JW'HWlM-ѕ orm϶=1}c<w'lxϟv/_Ow'8wgT֣rZiC!}[J:' O597L u/9av꓆3eʩ1Ogw%}.}F2{HzOhudϐpr~{LW_3YrƑ" bX-]vn =sdjvh6wǾ pfm:7@i.-B`iKbp.dh3ϙc+A Udlp] BW6c<8+FtJ.iCt%צ9~3t%h2.SݶTq;kW붳-Fв9v(tEӆsg0\]=W8tHn:~/~4]=%̧{>gɁ† ЕG?7NW2 ҕnpЕI= m6&;] JJW'HW)!>ont[+APFN\\7DW+J& ]#{t%(-+] ]yJ^"gޟyB BiwmX~R/w@&2ƀXYrT}/K˱$[rɪj8N=X%ɤz|f+ש|7?׽}Yo홽T,ʷq-f~<ʨ8 /w En/[H/[Խ `NXk F gc%tcؽ#,t Kޞ ȋvD+U*TL)K mr! F]eGZ4#Jэ5"]i ijSޞ Wf&75w% jEteܲ5tUF+xQnjCWUOω"N5?PnpőgrwCDW\9ȻtEw+վUOWMt5k]eh%k:]eJututŔJ0;npۣ2c }- XUy[*5wJttut;*"BZ5GgGtQ:G6r lyQ2ãIP >qn735+vfڞgv^Aku; ;.VT[rN\Ѥ{^k\K :؄n\9Chh!>2U5C BKDIy7qsC EtL \Au[*3gttutL;EtH]!\"hOW] ]i# %- بЕ60hh:]e 9E2Hզ+h{b2h%42Jթ?]gV=;S8~N_cwt\us %oI lb][ȃZ S[CWEdBwtutʼn&ʀ i ]UpM[ |btQN]"]RVG?et75tk}7Rvtut%(S0K?U3UI^owB;ew+υfe!6I4`\ LJ"x7u\,>S5lZ3kŒ:Mx$k`eƝޅY䂉 9'ؘkoI1"F54rh Mg&{9Mg:>AV &B2\BWVOdwtut:imɀU{UkH[ ƌӎN SUcv&#K[CW.o͘]F+H*:IDSF2޵,f L/ߞo)Qq糹է*/$$^pBsϼS{&0N?;ģ*_c_b;8= aF3h!\='hr>Y561E$h]H6'436Dͩ$Q+c^b/FL]YmnHnTz3Jx{G\ 2xau\Z-x1uϳ7bMz< cYX4X#q98)-S/?HSSߴ T?W8XAt&?Ǜs}J0^5c_2_2u/?n*~e\W6=R%>~K7w~r6C;e2t#o*}?RO/ķERK֕4j̎/SA1#1D)U!FA MtVa.tcL}Û 2!sR` U^Z&9 9~1$S 0 O5Zks R Jn '}Y>k#b߲ɢ/hjK̬QUw2sի7* 6}oWM*_x*cNcf(2+ gMҗr>|(/?FGoP%]ۓѣn,'HQJz )-&kr l- eM-T &qLLbͻFoì5?Gw9OJ_ӻ%,}SY$q{ ]N%Reza,^3{[JʜtXv't:}jbRez5W/a{#iDaST۱Ή؄3&` U^ A^-n'_?DPLxDE͙-*U[4"t"%"^Kk x5J$B$a."gA`%5VSEDjATA҇*&hfgBˤ&rT@ dZ_ "9q2m8]"qT[>Ob,%9φ֧2%U%dudj^[6[mg7?Fe }T-C)4z1\@(@PgZjoƋy2t3 {u~Yb&1$&'T#c!j?NTn(EN;D.oGVG |X9LMac_F53"cĭ[N}ҡsKE*0.P xb6RQ3\ԫHG_AKg\p^CQRţ=i[E3l53bmp2bU;Ų7&uf%"YNj/nU% 97K*4EA#I H2qA1ȥxXuˇP3~|x [FO+# l4x#YE .XHڂJk~ܰ9|{Bz; CI è@<,X|d*L+K1.06;BUb KzTG.HA2N@ZJ=\yO78f$Bxʚs< CU D8.DPL&X`jDzMԬk@ȉ5( ^c2r}־8 ?zJ FhGy,Re޼R+NpJ>hN# E@-Ru o&9WuƼl9Q.,;V1 m'Cr\t|W4WYۜ3w+FۍRԗ\ƢJ|*$V BH qq6sMvw@ɦ8u{bz]_Ǘ% CR2b]BY F'Jx S9 D ;d4I[W} n{ -!8cwmHLPx?X:u6T콭:J5E2Ҷ6~y!qHIMbK n4H )?ǁ -`F+=$y8CKhW 9HH=drhGcdCQwvq0[{ϝOZg,9~\QZx1Xp%7*!;%{!>q}g\4YסE ? "^kn`wxOdGx !t (ȋf_je"a}0 e^6 S((gAIObb/w Kw@hL+漶p5jfP|;;<'H]Ep-Z>'L 0JS/"Mr7[uש||V~(Bf<2x8p7+o|Y+~uaoCIƌ&ߏkϧɍiG٠qG(U~]>(~n@|5!Lr<4\"dYTƓ asg ߼B%G%3zQxUJq3X]}KX]1,hJ<%9Z~Wz ifɉøY8Fγu4~ۤ_KL?`ǟf>|>_2LT퍧g#[kj ѷ8iM]{E"O22 pU _p0 FAg sO/CO#G)rV|LFva4x6vamDEa5e1}LŨqkp6'P#[3hW0۝AYautY2]qQ*P_|UyLLIնU-j\~#w{PmThs㍴Ycjuպ#Tn"LP]ku1ob!ELT7=CsF ;,D$'" Z=SOl@;E* u9#T2K4 JHTQ BRnb hCdTkzщ82oyttW x; +bsɌ;6Pzg)CY$S|44=NBVQClmA\Inq˪qʇ-4NO1FL)-R؃@MiϝɇvqXq('$rM(o".w}'G{-_ŸJٱv@ٔ12j[˓jWGCJ9ōiU"bKRm=3 qH ÌМ]J늜}KxϮ' T[ϭ+uz<`_-f~j` r hA:嬂^}&e.:%vwL X17 sBȕ1' O0#`*#~,0@Z,:|ڐkH{㵱SJf@1m8j AHau2LƠcZFL&PFs+%c984N9FrETy-Hb, PC+ȁ$aγ\Y_WInSx%NDzl0fJ ӫ0)Yڜ=O.Ƈ "<kNCN=,rG!^? *6"gxǣl|1 XSE`' &+C"njP񼪟p•;J@ A*e I #)5BYŜat˲K^<\_?H_4rZ*hpݮWGďlM0ښ:yM;qW`3XnW"(#,q_mDjutD#Dh& 9$h#3-C8yKFP.T9JtvzLHH¥nFUjM3|vd/:H_V}>U@sly*iAFjD!5;g#hO1<7r94@"h@2lrD`p'93rvsFqRc,׭93En+N3ip*emHyQA̤^J@9ԧ/Ptӵr^]ƽ|#GI60*|_ DKx)YP_Z*K/Zex7 PC\{PŴ|ӜZC2 llhN0g?Vʷ_VqVuO`t|4l>)hmʸCql:[^{ui:ryڜoRhҹ!Vny^°H:!rF΍4Hᥔ\vC㙉pc6-Uig7DM%Gі0O{/GDD GPn;7 #$[8m-Jmc}J!SxMwvEnW1.,Wv kegL G)u~ԛr M# ^Truw0&XrcV)NR18CNw{-(]N8؄ -eVi (`A!Kȳy3rޅ 'N9N1X8}ĎZI3<Rǘ%:ve Px7" 6vf* p* XEd ڠ"+uEzPdnϷbZ.r׾! 0a@(Qd{bĽ$($+uoF;BzF Wd#jL684"//hb;My)JCHu,z%\eB ذ@G \81%NzesNu*5SgkL ؎R3cӝ/Ùg9::< \]wrS*wz9vߴ#Y:q OT C!fH ^ wE%KJ* {hM(cuu㾫Fx>!OWaTƗN'|(H/߆I. {{oշhiԂh ۗ(w$KQWAȫ66KX_+`n-jF[eilOH뙐GCZ\҂],vၱ+ea6(L iuB[ C:zG}#srfW& akӥ1PR&|+lBYt^`3GF[".*~o?r@(gq%fM+=Oױ3۹93#<:mCJިUܭ\q 1KA\va=CTH[?{׶FdZyȋczbݝW[kIɗ̿od(Q(QdR$lt%V8y"3D(|uj /kN{::t 3Q.)hM Ct2Lp'm(sYFqZBLԶ3Q{ ZO֏I޳VF}ü5fݭ25ܶUߨށ&nZf _*p: @b F+#ȍ5G}ѽx IU^J;<$pp,yo~yc dJ}+]N ?x/A9sr V(%c \0+f ])"ȠXÏGw#WQoоE>Dh[ sFT{<zv@E{9-c:{NjY&}֤ϚY-3xΙXFp< Rhn& #s"5ݶԓ>FӦ/s<&N:GDž 'xJ I:Xu3XපUfT kL+c6QvRdγDJ  .R͜ȱ13!"DK-D TR2;2C#9PG`$AX,#UWMT`;CTXedEeϚNY4L<20rO!egeѿ*̂uw}KWsc.B61 eb61w9 YGY w2}{sͩQ.?>JdGYJe8!*+8*NbWde ~Dib0򓁫bbWڅ1U@à?B죹D?z?эZdĽ.t*;  WQep`.K%fGe}z4p~qA U"˛;$+Pj_FvqH@g|.:XbؿKСiAS=Uo 7!*2 @w9}ln[ݱ=meK}v(AIeT9M 8X4Bj,h#g`+&*;X>W ҺYF,hbm2dA=ƘZi&8;$n!1C==,f8JMsS}6]zA8'bbZZKaq.Ly dlK}`MLcDҾMufu#d!(XD"z#(/m : OI1ō$#兏VdêNKe0xÉF\B"X f$wVyE@S9֠>GDCŖ?Σ:KiB#5V1A 1`89^FYgPqNZ.V0$ ^"64b1eaKdV&U0.xSZCy$fȁ &Y}a"HE/83F(NnxX͜OyLf[i-yeD ">E$G'<,Nr G<7ViVVư34Pn$IF "I%ŵL&ʤh4PUFjl}G \4YKEYEņ2VYRޖG)dЪ1e+D.II$B}Zޱ-BeX-7zn:" t e ;EفW(mB[)V.X !%;&Xx !~!s 7I"*Xtt蓁'C:Xc}iPdUfR6FCU=J曋g4~Z u{=; ϗ#+ X4Z$N>/ђG)-),!Xk-w:eFRˉ}Ru(B'(F<מ# U3 f]y)X +2}#¹Qzod's4FyK] .D}鬏.m;nZ46u;YteÔe> |k!t - ҖٛĤ>HPsX6̃ ȃJ@cly>S_\#5rEaB9>gC?NnHPLia,)&JI2huԙr!CIc@p29t *8OWGiO[߆}zpS5Oֱt(3Q]kS6OGt YF1aE×$bKi:vp! Dkh;~>>J:E8,6ܘKu0* X cd;/:駧ԃmjz>0\QvDD" *%ϐeu)hQ%BXp*[E. ]}^҃3/.|Ĝ+G??PF}F})?5(7S@Sbys1vHG_bIo/n?Wn9_tm')ݻM?^Dfi7syZv9 fK2$YV@Y?2WOz^YpWE/1BPI4tyJ{ cEpm޵EvMsqot a@z3mR}z?Z|>,h sy5`o+WtE?svӈqkS/fτ˳@?zDfoڭ1Wϳ?? 9[2qLK:ř[[zW:iA9^3ͭwјDItKr]":ȠUW$'LM ˶=*Vy/˃|})ϋ'>6R)J68prUEтb54RˬT7xd߇f)Ϸ'an[kR6sBϥ?#y|F[lg1h,|3D3?f {ƛO.+mfyG꧶Y{'9v^yuiDp^': dZ^vL1S-U1>3"e>eL+JgyY3-(o~ώ;R ŗ7nЀ`Z$x2\E )G=gJrp2htJ^G&ҪLM| Gz}~7:}W>O뱃xc5nSSyfqh 4rALy i)R>wi6)гtjxpWq321p}NsޢZL_Śݔ  =fԓHQ?nRds#+ ^x8|>6xk\39&GJUK:d$W, ³d Ɵ-3dY\V,D;lsF@GS\7nT| 0dWDQhhR¥J ?n^GM7j✘դx~_(TdE7{ mK!˛0T~qHgE2nr~HǫY)czGr~+žJlMw/%12\{W Η_v\/L_N/.nW§ ސJUη[&`޵u#r~0I)M& lj|Ȓ*IE$:"JCrf8VUkp K Z<%~.\+4tpa,q#ٳ"N?כ$G13d?>tmvvߩi3\}ue Cyg6ƺM&ee l ެ`|TL4 wxCԤ |{] 6_~5}`r|-R۱a׷Ն mޮp#x_j]$U U\ Tڼ7$7uv[@ȴ\&ҁYQWM|iD$'"ڰ Z=P Ol@;E* u9#T2K4 JHTQ BRnbJBfCdTkzъzz`6mґS O@v8~V瀓w1mMSjHX'%h 6iz ^+B/@/ 1E63WpE\x3N>V\i-B؃@Og(;;hケD3AQNH9cZG Q+_9^0n1A Ɵ`NV0v"9rObY81)R.RNq, ,1waؒ@[ rR0#4'(xAc5rqf/_}v5xZ*^lu&O ,㗃ߎ z=h `9 W jxOj.Ao(gDXJs^1pkik?Yy;$R ;cB! "zረ5\ch@h/ڽDP0T,xaR[ERt"'pa;[#g3\ҵ70N9rp#v"O"ƞ 偘(8  !rae `MlBK+mBmbGdTaXK"2 )d}R=jʝX&9N'cAӘcAUN&P# Q[r;DӞIς^)Î0WٚI{%h} zm+yjIsX>M\j: ,YOqgK*Qk◲,\| oo{{/^0Qg<~5:C7 ?E{po]˶ꮩb9j·闵9~W~ {k@g?\?O3kq[f1ׯT m9tܥi]B(VG&>JO"z}''XȬ%8 }+-7a g3*4F>`ؼ"Q`'}jC-{{&a(j@<gQ5H\& tSJQv5y܆UGםq!RvV]wP|Z/`؂@F`cnR9 !D[1eWpB;qGڑN8oBꔼkCbTkH{㵱SJfTBV1"hD!e!x0Y1c"^ˈiDk4 cD[#3bO#-Rq "}Ò:8<[@[_ XhEAP ?ty[Rֈq6[4t:qO.@is*"iГ j 5!c$ W܋h[3v4 Y5-ބ0[_Lt `'l0kU ݌ '\Yz0?-n3º_0dk* "B(ק݉I;)Ƞu|?គn]nu; |n,֛|އ4mWb8[\V7A n9{|7DF/7$"k˦qJ7FCZʌ^|lZ9nJqd.IhcƧyYŌo{7r^ˋmdͨ[^~=LF5ե3 T~5y;Bnh?pڐQ9v/zum7òn2YӱPYmϒntHe;H6@jҶ]֮J.>|`2J˕>g DQF=Z.*>` ,q1y*䄎&Hi$A{9o)Ԙk;J!XhG  BXYL^j("b1h#2&"ݲ%5rBXd!%#%͡vhv ݆zvY*S?V|L- UJ1LBkw G#XbLyn rhE 6&=d) F刊`) NPL2mkg8n^7 i)JUY8jY2A‡dJKI?7YuV,I줌'/uãOn#X'O pד~QKA}5^TRd 32(Pzlp!|VlșgITN*x62\VqhU.\<ԃ;YvmBwP>}m~> $WݥSgA>>Bukhha0ϠS؁0h ~q>_PKj8^9JՏsZ5] ,-inCl0m 0ENiͭ$G~ZeP91^~{}߀G4b)u^i$4,( 3;`)#9f0‡V) (7B?{H]>&C@>$Iv$Y$@fyIZ˒Ze{0}/K%Yre5mLh=<}$)N^'AU qY#@=H{1PN{`v|B] <1wGU;svDTXs9|Ig:H6=)I}đhBQrju? A1IQ] Q9UoEnZOSR$H4 /F4׉Dfx35z EPzBZg(׋JkIIH\A"' 40 2KmК3LI\"jO8KNPIP&A}?5=}qP{dh PUg_8M2z>6H/8Q (nU킕 e5pDaI){-"-{{{ 44ηo&{0x)fny^٧߈o0ϛsɗI1*LaP**ce'H ӡ4i,8ALf!*$Ŕ1 *,05EY T8Ϳ!*z+w+7"t$E:IVW١ř?ֺ jTkSWp \X7nl.?$,xVEšy+BXԞLDX'ae 4E7LP5幍C7^(j4H+d^`9(җ1xzw@vh(/A-59jp9@RMn PxeQL縥PTP ڲ\*d!dpIS(< f\rcPS2x‚Zl:5^GȟeBb:u}[BO/ĊᏫٿ!pɬ6@KZפ=1Hef":"2:'F:<LH2!Yς$HQR UM02*Ű8 )g8wyds m틿hPxiq H-M4c˲ Ju(6H#$vk'39B4΁0D.IPυb#PT\P XF\ȵ(**,̮GM%Y$K'*a?L%cȸlgC)l:q~hؠNnhYIf17p}#*͔ecJ&%EyY9b vF^I5g| vtXy˾w*lT5~/<$iA Q$\:93,9BE!s<:dS~|n֙R{$?~NI(tAﬡhulθ.ۈ^F7kC5x9oDU/D>'ɬPTP EemR:9qPڍgD:ωtAQPgP* ^1DG([d#^j s1ɩ4R D̳\ -SN8e%#1(J;ar6z{3[~>\{GxϜ)\sz/rK~(%s~I`r.ILW_N|)ُ7KVm햄++8Ls/=i>vrG~!8/rkP'%4cvFX:un@uLA dWWFtd;`K~Eeiͽ}9eL+"<J< ey[cşNQHNpFEm{DxD{=0 4#Zh4T"MFP͑1bF(NG" ||F~52y:ƒcz[b7mƎݪc08kf:6!&w7nw~(Rs. ?N#?dG!^ƫ)&nE$&?Nwg7FKw&UoWT$OV?)ieǣr}WqLow@u=b,;* ηy/şqkr kp oxz}yFeOM\ b+Nn ac͞O?(ji6lݿd_S.ݍf፛]Qݬ~m3۳z4 ß`pkN!h \{&Q" h>M.jlH=cO39HAPD.t~z6|t3=ms.Ѭd/>l>tG9^<4J7h1QJ8>gT -!tu_v)I{ņpWR CWbѶv7dG,1KuT?5xGCc CƟ2137,w>>-oxWeխTR҇y2mHaEv,k۠Y`]dJcU-o'[9CC~{FꙧZ #*̨V6&>7<ߓ.7l>ovhVYxر@m@EUg&'mԋmbe%t>hiymlx)k_I0DS Q6$׎Ck(mS&YebtZZR[Hw͆Bv'WIs.]SΘ('O"XGA$Z 3*'& RX '{gI(.4.Xg<27)XJW.6#Wh]癌#WaWewd[Q9w|Lvx-=;V'$~j,T` iW W_jkh8|Hw^`Y 9 \8xm?YqGH_{W񊇓H5J^lY%DO{{Hq0t3Yc ʘ=UdEoec~y:s o޼gӯoϾ u/޿Ο`a8nA=Z?F._=-SY}["hoBKŴI|B`@k=WHh$I6V82~ HJˍv+@ Wd#jLu~Sj1Ķ RE 86,J˄`G \qcJ^)ʺ:|٫F͉ïם/xaUuY`lNfb`rBȕ1'ʩ ͎Ш=B#k;=E{iKYگB S Q)g!!XzO C$8p0JA<UX-ԂwfƌP{-#c2FlM4N8Fr58Z(RmkAD8%3tpypkȁ&T5!ZB[ Be\g{}7ۨ#t^?;' TJ^K}$T")|h|8j 5!cZf|˟If(|̪a&\ 5X{cR0F#s>a{)?}-kxه}t> 3Nz;w _/A:+B^-E!ğH+2M>&`y.{WW)<-p@xEȦ6尠Z,>ð|y ?͆X\mwyZ?,Qro2]^l˛ [C6+E+Y8tInDKbKUu}:/nra3|vl]5Z]Hܕp0X^}v=EC577 zy뚆KC7Qcޕ^a:5ǯ/Fs^ۦ@2'W*mq.L6aF6cڬ4gq{;l!e^Y+},c \ E^l=hw$n%#ݡ 9 sIО)F$f[p"đs\bGr:khaW O@d+Kee#d|QՔ݄}d/nOjr-2Z%4Hԑ($F{笰xD9 +Ɣ@O Сc&LJgQ9b0XJBWpk_KqWMzٞt%γ|&ϒҊ}YL^5xp%B8zģsc)}$Rx)%!K{]#diؔNB! "zረ5\c\Г@h CQ{0rHR@iZ*e;Q!(8v0)&8 nH%z Yo/S){( ϽS VK"'*{Y8JHH 9M= Jٟbp6ȓg9ny &8!0)cǖ"Wf%%~H :q)L*!dG(NjRt Rn::\cAX vaUN&P# Q[s*HN@DxY+eXh CƼ^k/8a5Z d ݄GUaxt0`r Tnn981Rx€/sFP[tt8Fڴ33E <7)Xc {x^=2VjP mژj &05avZPBE;sbaP DQ@U :(PAhjfhªq(|; N^lM/] /=PH٪ƁsDA"".4# J99.˞Dcx6&VZz6 'HF!e+Drq0 $zFE_+/E`8.&Ⱦq9}%PG[sBz`~)g0=7kW085i#6GH+R3Eyio$%Ǎ5~L|}2eE6Vهizf\n0fQmFik]6W*`k ;ĬW $#Z{y8^s [ *ۈ|he(`Ir-iVʑ&6dd|v{K2`|(q- \hAoz?7 $2n ԛXn*9^ÉʭX.Mp\^1r5k5i=Ғ )ҪM`z:-I7M6 H7?_5%f~a(zpp5=N`Fw.O*Mq O?sWL#,8ZIz|86pp}1f(lQq\,Lmm`;LPO3ȮQvp*O0D 9C`/lNxyw}N\Jvv/3eeKť#-Qʨ} UsiunqӄﻑBK|>CcrlRc\qaHhQ֣$(Iоo=JRjm=z[XRW 0`UWCQWIڹj{Ju+&,ᠫ$.DJRN]Du Vsj΃E-BWNsoONlEm}__ R ~9&9§TtLRL':{RCr\N⪃Q\(kx5IJS/PM 0@`$.crUrW^RuH ~ߖJrq(*I+CRu/Q]) ?O I`}8@\!I$%azJKB X_k'ž*&6Qp(mxUnpihnটϞ Yc*v?IƞG3hn#qV{O1tw1O89=.ƲXW0uͥԿ:je"a}0`LjCZ$:P W(bњ%<}T>KӠZ" ¿)<%foѧ5fߏ7W[~?z8pzhF bܧx W[5Ę14ʪxzU) KEM&~ b]w^&]ؖO"P1$ ׆Ĩ3֐,kcV\QŒ ^vq>B*C ՘ID>zﱉT iISZgCʅf&$S^p…LJh@) JK:eWIZB?WÑr*2Zs" fc@6{&%d)mpMyYwSjh>%DsxDsS婡zkʔ46EթNz0uVot,oot7Waj SasqHs RnQ"*6)]muөO [\LxpuWϟ˫W7itzL˞]=EpeWlݼUBePA7cbɺ2nfWA)ʺ\sfmJDe.HH7NFlG]n.דl}a vZ4 |T@J)Dd$!1 KX8ead24_ )R$a5#X| ðV[DR92tF"!}ɾȹHŌmCs6 ӧ&]aǕ!;t0\LӺ黭b K=|].QfbH[ԉb`)f@tv:PA$/zabbyTh-{7Y*[ <cPV-HB^,+sǰɗ )#)A tF*d5y@)AfQ :ɜ!X-`¡|Yuu,8֡0)$] IJNY‡D$rt6D ~Q\=G%cP]#.p%dLV(|h䒧5_RDd=2sKrO0{eT.󮌝yMe8h#ToYh|xf-~?DO=;zX 5(*rUeK|i~w5<=R-}5N5ƿaߺ3/?\x~ ƣdp\Fj9W$j/&_rrGy X3 D pHo ӵX9UfyɂHy-&zv9|Y֎Q/nu Z)uV}BFɴx1cU5|,`%CPI送F NHdU@i'=taVz>cRy\3udr:S1gPP"q H HRj,s:=ٷ@6O\5;F9 `ݍ9Y>y._zsVRw'uz=ugQhNxt4u)G4Z@J00+2m}H[f}|!Au ".ƭd]ksu,{CmWe A;'ey᳐AK{2‘8C Btq-TE땃]2YW7?]!y4Qc&ep{EԔ>-[ѵZR:z$tdšL&GltwZ02I97洏%3 ɒ"D) UL͇WLKQҷ ـErҘ(5  xuYKE/A6C$jHԗC6= 2>v?TkȬk%5 Ywvmμ\웏0] $gTĆEl)Ro|tjWpo>RImy]7 <7?/w:&7xX1؜6뮧(?L(vdl]*c.F &H[rByeUT^;1D+@3!d(ǫ57{AϢC?Io=08mk]yғ:' ~U~o9MT@~_Í?=K=bȇ~o>tsubG8f.{vs2#4SVGĜXrk,}gۂ'O;xsZ{YF$UU::n_mgv?/`<_juâ~w?k-{_ۨbŇZtA%b:-'4Z\t{v|g-̍] le'whM}<0=^w6-} VSt#F[UECCCƾ*5—$K) 1G)IFQVzY$aɡSD aG_Pky'tP, @2LhC'/lA H^s< e`[YK6!dBDW,#r\dY@z#kd_kǙ?z##ֈ @["iAR&]B4E!:`ZbMJ² a :$2yK M2nYN$R4 E쩴&6HEW MP~6zc}\:¦a8<;K`&u`4}OF9"61DtwPIq7iDAu/ Mb0)d1/*F5_Bdz#v$sS F-Nf½ ; g<߯\i\#[mqAkxAg'k DBcd-vUz@СC m +} ! FQKmビv cIN6l.$:m -V EDiDȔ*dH}ȹNN j'SVw,/0y?nR9w)4oԛoRMt$Ĭk˝ #haHfUȣHYŲk®1`ۋwRd6jz-!eP^5{-G|=7(]{^ώRW*Gxb{ V؋؀@P|RθT^nvJLݷ ҭflG[ҿ',1Lb)| lRڊ`[\ R6hQ^ܬkzТ-֨Rwnsxp?f(Xٺ_|$B*?MQy-YsMU| c]΍FdTPIxW7ToSѾ%{:dDM`9YR J.bJF:Ph+H^&'zE6儋N2RGJ!Fn!KUtѰ;7rֽQe U_{RYo?';6*HlI(] wk3%%'H9YA.e:UmO. K@)cqL`AYG9$SBbF{j)Z+\z_.t&0,en8/YݽVYQϧֺkrj6's1O{㕂79?w˓?>-X{ͺȗ翕,^틫54*oT!C[fD亐IմY@{ՌBTbtqJmX!'Z*>* \)"#R"Jo9kJo7R*ѳ.]S]EEkVֻ gfiީcw4a4| g߹.V.>֦s6ex-CMЖ)!X&RxH ҪMml2 UlaN8_ėH?{ܶe J_61~?Tښ5xRI_f~A"e HA2!$H"Q`smUGSpgl;nlvzlM\c"3Y!1J<>dDqL94(f\H,BeXю \D98 >0g#FuӮm͇ys~572"nqψ=#>iycA+#,cC7K g l!m I G #$$8pFΠ&%#YS, A94afp 0xqMg͒My6̋ŞTx)2{[]`0X` j-'eʱ&3^dH ^lNljvlʇa>d(lyuYoYgO;ͮvNpぼY|Ӹv-na( 'c.|yG 8B#kp4eVdTHmYwQb9nGHZWr#,;B*PR"lp0-*im̑0QFU;Kv`8M;#ś_<8 :jobg gsXe )͕EQI9Q)IV!)]yEq@?ox,,Uf$ Us9#Mn7n(+U:cڀ\24`NyX#Յ\l -06rYͬ i}|I2 ͬqQ/ ;+ό)܅sU*g. PWPZz>{M\eBp3^K/3=15}09i8.i\\SsmPֿBp ŝ`U7dtI$Yq?0#⁛.㍃s"炁KgaրKG^Ojdp5|0;~ /1$lvhK wv!C)h`X92( :uo,Ub|.N^>% |E,pUݪZ%mó[|xz6eUL׹#ie8ˌ'ߢo f1?G|Z)AҌp̂5FտӅCx8p -p55êzWqQNnUSZYAckeWE zeWQUfG﵇E](ݴxYx.j.O/Pu_Y&.>HשHl7¼^+FHkYUj)RYQ$W0CӇ.M)3kb\Ow h]ASOg"h/ g|[Et>MFU?-Z ypa*Q .!LW!hU냐EZAȎ0U2! UKdW*tQR`w"`:CW]+UDIuOWGHW3w`ǝ|ؓy;Tc㔼S訹EH Kƣ`ʏmڅS`LaX t>R-cc\خab š]"87׌_׻Z@(>.,Pјw+sOSOLVw"鷑ϼJ%MbiDPj0#᧷k%j" 71pry,era9+f9*)~7:r8U$` XHTgS*Pϕj8pˑV03RJmʌ6>Kɼ#-!֥uC]ȇ +.Γ76˚yoXύ  E4yLƳFn789VH_onʹRϯkB 0yo%~$oO &?nJE;"`-:s}60yGDp;" x; wx㼳)Z"ޙ߀J{:FR.m +_0ƺ+thm+@%J#.E|u3tUDx PұoOW]+Qs)h-jCصp=P5JtЃ"]`%pg*U+t ]dnOWei x`N[pv`up ]EtQjR.Uigb ]E%f#+F>Jst睡зC[%G1'|˴M><8 uN5S%FVnz]u[|DW4#J&z>BDN&Rt"bhj;]Jګʣ+I4ǺCtw"ZzuQJҕ`t.+[rhl;]Ew8FҔ#nst1+u=6mGD{urJzy(Cd mr =]m:L:DWJp ]0"ʶyt"L);DWp*՝QWVv(^RRw698rjɍƣ<(0 PHnfQy 0t'ӋYn]*Dy YRmKw4'F/7=Ld޻My&1miLwmfiS@&S$l*BYR%O5(XIOok?cJ|x8(a(LK(rI}Q'{ޅ;yyu#g?Npq\A&1+,N oysc`?aN BÎ@|3¯P_鏧ɻ|0aOz;-R֕&nyhr$\Lz'_V,{ϳRz0K6y'[घT2y9#.C2Lg8KiC :ݜ8VBvHcY:"\"Z"ۮ"Jz5vjQKiV1J5sZl5&ȞU[y\:r eU?mAI_թ-WK0X1TX= feՉ+VhWį/Xά5J]i<7\\535e%^q,*7T5o7C?Y.P_l P'cTR䓳.?)vݦT& \ċqaJ$ѸXZѧهU};6%Njž's򒔍Tw,Ӝlq^F8in8 XqWG5Vus}eWw s:g3u^4/l(G%׹zm~ۅ\UB Rn8|Lɉ733OxtDWT ?;-Dll;i^\n$uhf#D zxFB6M~zhS+S;E rqx+;3t3%nX Nմg0]K|:7: sdm"3#WEXqeX$,yH/?aaeˑsj$HsWVS}G{W?=Qa昷;e=҃ ]x(ۗ#twpv8kMD0F auCN##&c;*3|a:Au#{ N9m!ڌ; yi"e^zvv̝{Z-^֣ui,-pdjJ-]JP&] }t70g VCW@9t!]%N3xKW̴3JP&x9U]׭fJІxQPCWC^H?]|;Nt\~6=Ƌ;6^L;URzۘ++=3vCPzt%_]dٵІç+!] [] \k+AN^+A#]Q!8OԾNWJW ]9Y{bXtJ\l;;zk݊hд )-|,(W~4φ>+uսׁQ*]O6|1}z̻r^x{MPo'x 46QY5}$y<6,~!o?ʏrvvTԈ&coǻhE9;:@l33fӴ!ޑNm+·{|szO/Ng_bJoOɺ/>w3Dk̔-q]P8R_o+Yu[*mlc^Ȯj[?ft<<!<,j牻̔w G346GCn^|r+8ח/ |R(LxBr]\37u6Sl?囯]W7F]^ov诋}oRa7md;giM%xcLkO6dgS`wkט]-l\EҼ76\Znz-h:?ٴ1zg?,4lNqL&c]Lj`P\Xm\RNָDOq2"T.dkoluhFqZ?-Kcu]|Ξv d [HX3֫g0֚H0_ˆðT13gIG>KwA019fhg+YjHyZ#{pGD[3er M="}mDfQv1G408[Pa f}|)60ИU>g EF94w&)ᑒ9fF# 4@~S.~WJlI1f] -mY(`\KT+gL|cXs?o.B2fIzieϙTҊm3}Nf2f8c[ulNÖmݒydk;ŖᩤC+`##;$v3 l,(-9jH) AtO!i0rk=N`1,[׼|k>$Ї4]zy;aYgJD+MlV/M8r`)Vg`򀢳|#h-y m~,+e^ AR:zCܥ+e $xtyMp5XLčd+BfmkK4xH QD&-5!")7fٰY0:TkZ Od^FGG|*֖ Ŝ>Nx XSh3to@:D%0V#5 #  e@lp$ټO +QJ՞ꇱN_4~k*MC̈́+ "XmU~%ٕ};\m'fH57]_G2P(SP|('X!@YPѡFidydМi"ϧԦ[ku s<{`&18ѠZJ u(QXP%@ eb(f !X!joꩡLEw` %R|ʾ BH{Y2H( ':+BREA+Rq&-RWQvUuA4"Fߎֹ"y̼:RF5A_T8z+(%w[ enieFP>6i1bb,dPg&!B/7nn{".M7YV&e$''cE5#&ZBC~ Xpعmgv^O}Zk*{ì̏϶3& m3Ska&a@AyD *}eU$3Bk] $Et<9!. ,J芸`&) >@ E&rZ'd^0P>DO:#./1/z@ }Hd 7x$;1-^,ԀGwf>oɍXż f@5[#e v:"}w oKb%6L 1f,M&XuOXk tdDrHc.m,56X!/Z@mDebˌ +& 8gk ) 1E7"<{BܶKAEKշbhǎ@H-Do@wHBB.%m)jPFQOR@3dj gB $"2Б2B`2+I@B4# [. u"NQ0*":jѝXT.UT4Š|;,⬒)ئh{jj-[H ket=i; j@e֊o;JP)NՌ*"2^I؍@axT_|Vu3ra(у62؈q^y⦗.Oo`xsr.KeR8= B=Dtt 38_؂<4z16`Z ]LiVn-ɵ֒u"e۽6aPN zjy4B}۽ʌ<&YAIpd>kG6G=Ԡ ʍ::+CuR"˵2oTP=`G(mEAJ@r-WdsA)Ok|l@շ߬'p6 V$–%>-Nnb!G) u7ņ(V1Z8Qr9)j ꠀ:JhFY/2Ru`\3|ick0ٌ+j hNp> 3ZVjr2vJ4YӃ-t|]e^֣^AH[ջ:H{ގʼn{7/W0.q W) z!US`Qܢ__ n8+N2 H̛~$Poa"W#6O?~I`*Q+ 7zu2Gh(?apzoHDU(V |ř*p3{dQqJNNF>GQid j?q$F-i'r+6dq̇A~A}fRq}%Uu1WX GpN-c|I?Fnܯ 2G?4vջozd~h1_0rSgdEYE􈊱dcA.86O?Fp -`bB WΌh;CZ9p(iDWWb֙}4 tutceIt))'RZ ]!ZfNWrH9K2VpY:+ʡ++H)thUיs+k5ʱt +]}-5{$5zBW6\yMZ~=hߤp5;)]Ci%J/bpZ?;SNa`k]a-y)th}+@5 e)8CWc]!Z&NWΐ8*DZCWc]!Z{ P6W[:H +().St()PeCW=BT# 3[[1, 5ꊦr`D+hiQJ54p CWR ъ[Rΐ Lhi WC'r-/1-t({hc8pO"=uzl+xS9Pʞ-1tEzSne6Ϟ]!Z{Bbs+& BZCWR 't() ItIYWb+De QځΑLCWE7wEWVTv3+Ild]xvS /' VwA"i3_KPc|IlF~rj7ʪ2r1Gﴱ%Xv[ay5L6pQPhrn}-H@̖ \pp=v(k+-$% `UN15$g|<](s+#%@CWh =]!J6XWHWV dAtµXWVǮ]b= Jӟ'=u Vp9qvhOUBJ޳i]:T.DAt5B!]!]1-RCWWR 2"NWΑB Kʊ+ˋ+DL QZ1ҕ0x>SAt=umvh;]!J3ĮΑ$Jʾfӊ?C兲LsS^5 b+X)th;]JJ@WgHWHeAt-)!Bt(v@W/BWJJrTռBV%cf3+I BZCWCWV[WR tutō'ժcQ7'0(|8PȩrV~mskBA{}hXj`j5J'Ks2\(ٻ LK:^?ME%Qkc%:zرۙ1(6Z;(yƵ[mo۷oZ9hIP3u7IVB\q*íKX|?-߶s$^_D7'볛]q䞟HR/uqvЁyku j[_)>|?͠e7+Ů1O@T`gAcƟ&S}ƛfx5_RSv~qLXL2&|Χd97bYāi$ II Y&Pɮ޸/ȚTtTrTxῗWI5&뀛O)Dp`%ΓxKeL$^89'`=1 f!ȡ[q@ 6kCRgTf6 CzM\#8|Hxgڵs߂er׵;FPF¢_(X%|Z {[7-._O7⮡?6M}KsAt+N)y}d\HS]U#m49t41=iO-L)}H͓'b(ıȥM (p8a0:3j:ISuj\jU2uA鮝S͡\3'met=`E3Z%ђSfJ4ך4j3#y3G'ıL#i^y7Via2s^N&g̘37_r>kD?_lT0K4[MV_.v_mf:`+&ƣπV%hetlaf=%Klh5]6[2}ĎF0Plc%Ft={:[3.YgS1+cFǂ~DᲣdjY^\Q* XU4U^IYqfT=k*IEyHE^9lT|F&5g3 GL2ha4ASڧKKb^ 3xeR11i5Li>jK3Qtmvq.)*_5wH|#H9:W: Vqa%YG5R@-.6)&hkL0W*σ#6YnQʛLiҙM4&Ƞ Fu i W/eQg,QD [TG#SIVHyrN#L0\z1+Š,ϡ,|t|:唦ke!*k<"WA1wV ^H%(ɂx2ΣQ"2_''7YUW( `U_ H0;Q`(ci灹;Epl?iiCELug_:0s`J,O1nޒѧjt3g8GPvc?#on=<ԃ~>PI㱻.DsXbeomT4Dg{KyU,}r_RcwLc>pU h*F+xO0#TOKY>{7m~64cq ?0-vzW?C7o?OJ)r"wwp;Cr!l{]uד A?ӻo寳W60LmdfΐWl[3끉OuKKMu|­*i:k*)߭r֖v2 2'ˡB # 1Vñ|ҢSuУ{4 [с$KH4.+0d3nDN<И:i2rѳ9'x>!&3׏>l72xvtc/9T˂I]PC̀z2ډ],;9щ#*`9FSXQ sh{cDP +=N!iUпtKJhiM \VĽ"&霩G#,Jrk9n_K+Iw><.3h쁉;߼Cz ?x'LK/?z96e1ge=?j"zr]UuZx*p?y^kv^Ng%3p\yTJp>295>&6\0)93K4bvIdzEΚ3ò#D|hnf''LA;Q^L&/^ T}VtJS{P!'I#Ch-Kir:Ȱvz4螜 fEŗ,~ovT:T'2V< mY)m#IE$YI.ECj[1Mʤ=9O%YCTSYzaOs穮3w d0N;t@uLc{YyfEB*1;瓯qęI#aEqD f Qрr嶽&|,Pb1S-*d !`QU,GfytƔ7{`Uw+^1#w݋EYϔ*{Kp>C} 6ʼ:]^J~l5cϏtnʏ}߷!~].?5$tAJRPFIM֑׾2$b 4xeCȖMЂqEcc Ruih ֶ0g㩈,sj2fﵲXM*҃2jU [kfyX1&.k5r g~߸dݳxrKkp^?Y8o@O儹boIo'w3c27 T7y:H3?,Γc4Uvo݅*cޤb.ØSCx~~lyݱ#^4O]Ȭ Jc(RL$2DM^(`20KBݭ䬛6l1BwzLl1&َBy Jm`5O^k5(3QLǪ;OYMU>.UJLAu:cbhp)0 8fʌuWvKHB`q#ϊSI:Et2xB f^ G(0 NOlQAio#5dLn IUD}f X[Aci2t #HB?mᏏک$ Z`  ye\d{*RQb^'FDt [@H]LID2F╶:Vj`3x/uj:*]q{LK?<ϯF9j9|<>sw~3ud3y|cOyM^5p!7jE>RZÛ5!k"cԈ+|@`(h.sҾ !ڎ a`}.TNC)&t:IiM* RtFv\6[b0NXxT,Qm*3Y_w|=j>\ ]V9b/ j>CL?1QTc@"g9w3AVm:.oC{)+V6 8v"ĢL$|A*}hj:#6, j7[#AxjO v`cƕl.-!Vo(Pڔ$x)8fs[[dڊI٦xX 7J*5yƚLzU@đ1?HYTTxl:a/귱nǂǡ(#qBĝ-'`COcjsgIKo"dida"$#ZyRQ۶ 1)1Ye,|%Q d`ȑ^S@d%K^cDl6=?j`\\u*9uc\T.N +>I$?e%XDJi+ѥ+Hv81pq4<8Mc<S~\\λe5?4!3~I߾;t`u19BYLG(0B4) 2",smŦ0N,-(I{D1YFO]eDQZ|sߧ7'CyY"ŧǖUbD6F{ElѮd!,k.%oۆ: ׬kiU$7;0 U Nw5Rij/t_unm tˤ]I2nެ{x\?U0JT TDAѺ tL;h~OV&- SxHJ$(d} & |uL&0b ¤ˏҞt4c,1;<~]~{.jhi>\Ƨ2 0 U"LYUTuu俗$"V+|(PfO03/i%G^$N$K0kf \b( $c4<&_2Lp?wNGzb&腾x[;l3/wb*E' d3+) Z-uhj-d%i'30G ,=3!wGy/skΧ)ՑRBL.n lÓw?w8ό/ЋN!zWȀeswl뼈SΊڄ63%3 Pf(>)PYdzZ qdI})j $ ^ xsCntxD!eR('Õ":#52- xF8!iU͞ӭ*gZ6 w8fl^\AAlիpZ^,W56E͘0G]:EFד?Ӽo1m7 6MG7矘>eIi6t~,7Pxkxɯ<[Ǘԇ?tvqK~}Aݢ#V'+՛ ?ݾWz}QED>?ߴLwX?K[pRf̜7ݕOTeqcVLy y3xtΗg4_koN<6K7hsTnF$o^&a]/Yq\N[ _'«1ymڼ'Bډ}Fio@EMX%@c͚JM*F"Oޗ&NȢ-hˬ+qwdP!%VC(ZJ!@8@:\R)Ƞ׸Vt|,ÚC;DŽti GU/׎/)Kyjn%݃=۳6هϳ hR;ct1%%gC Ĕ wXM( y9b6Ԯ()b@K:HijiPirYj!Vݯl`km[KR1Pm̰ KeU."d OEp$.K䶚>=.ηTĸyx5 _rA?5h>N~Zָ&0=iEb&L|UI5_ZoY%^&"\,Xu\fW o?Yӛ·?X^=nV}\u9|Z_~6y5l -58?|YL> ]>Wz%dCW=k;1:ڍw5X- |RYtF%кE~ۘU41kRsѮ֤eAlsLƔJdJEP&+]9;!EhH֥b i™daM19(JHkaL%`,ED!+kuZYYRVLn6=foJcL'jHf39CŠ_HoRD/P}ܛ )N kt6g HhbU >`Y#0u)7GeH*6[mKds,꬜3uV(ѰzlVcN  f!!ãeۗ]vgX  Ƞ/>*٥l0_IJl䴭S L[qX[zw"=H#AIN] ꜈?EW^%1J)\eOlVd&"%(}r"&Sza,Ljs,1R@f !j1U3d)\(QkU{T'FY pvd5ftk]{~`Ѡli |:?{ƑB.6#% Meq X'XbCxHlkݷj/Kđ(s;-NsׯW;W# L3i6l*M*|Q½՟ia:xzq{asΛx;$wE4!`"oG`&eK&.ѩRrY4] E(bl,yU.pVQ]E-Z2T3|xaI2z@;,}k݀|7wpڬI.N׫weٜ^I Gq_4>h-頜M*pfEU^ZB,3vzQz&DqxفBJhX BVhw%(efͼsk H)ֻHLli{8QپӭHuf;#Yvkf񢪬-}"·&OU?mڙڈ8@[LJXf.ˀ@Cb1:SqYEDC&hޱ3%Vy28(X%]0V䦒t(D֖<8:` >O1c0 GU[ւr篦I,9Վgztc:1 x" 4b ^8%S!#PRZhGy hsNJIwc ~Cc&`-ѧh⦣[o9ZK=˪H"Ho{}M/ݩ}ןm4߶>ϴi GJ~͗ghεks?~`&NDz EHR`{889`C^\'f*jY]哉JV"PSqR^JiKRV AF#@PJrbơ tɬs;Ζ&dCr X'`Ywf| 5;klu- !1(ŃJ 7Ne)*mD.eЁȤbXQ:C@P'ǹvs$#Uxl)l<7FUjQU%v+[vWЎi^mn|m:WmS4C~dCAFXxµ6H&|םy4[L3ȴ?gG_b8D02+Ϝj\\!㺗+VɺB`F+5,1m ) ȕ@""Z(Q޺:DrJ8`Օ|+5٬s{\!%#WjˢWGL)i0}ErvGQ:-R-Jrآd$W\Y`;\s+rE]o+9Gwk7+Jt/W(WRK)s+6\g]++mjr+<+ :?wE` Jqu}5Y]샩0Vb4!5waF5j=Qts#EMCO姏/3`7ER.>*R/ U >}Z7aqOG eYNe4+>E'y}+1zmGRPsqLU}`z=0 ܿE=k +mՏiF_÷/B颴Zt(l*C!uVD /&jus~CR4P&Փ[3jj\^U%_CU@\TpE%%`}Z)^vR[v˒Aq+t| U,yJ*M_[dwY.&|$3 /«xGuy]: . K}ĕH)-ҭ< 6(S D)M?pS p)MFrEFf#Wke.rEu]鸗/"Wƀ9YW2Z!|nVv^Rrure]w}"Y6ъo:&JיQ压Y%JT.rERr[W_\-^1`0hgjMǭhۓ\-\^[ܒy\!,"\#s+בVrur%e$W+ą}oG rErur%ldh.Ȯl i0]+T8d$Wlt6r $WH+D箈r̓n/W#WO8r|Ub!Zdڋ!A2x[& RJV8\$ds-*Ԛ4,VU1<ʂJP*(|\$&ǫ0$0XpGL䦃$ۧB@`d6rB Z)n!%C S2 B6rE4+ˬ*#B`\bu"JϿ\9ά;ɳ+ehMWұ^2[9Ÿ d{j~-U;J1ʴ+csd$W+5c."} ԎkrEJ\f$W Bu+X/W(WRu7ݗ+޻vr+)p.WDM/W(W &+Ბ+•:"Z%.WDy/W(WKvH|J{t=Σv%cY6w-S€ohY{e/jUc}VSy)W*Y軖onZMa?.o/~>ijY&nϋl<ʲ htZuҖAJ6xu\җ1[)6;b]sI/_qWc]Fwyoy[RMjVEIoZsW]~c-nW-;G(E Ɋ~Dlhݬ8(5G:{ sBC /09Bi3 i*br\^Mez%/VR eĵ+E:8ВăԼԪs]r/Yk(o:Wz6G $5}7h't8h♟x_& j<)>-pkQy^2xd` +)RCAqϔфYZq=*~q"&ɢ cr_J|dط %!DQF)*#ʰ}ݔ\ )h4;W5J2F0b+sXqb[+/|7+H4{ iY΄KZO k]JKIXba8$.0љ*r_w*hHYksVr|qq5BjS]*ԦRsf J%DJT2Tcs&jI>-CQ+lXcmFRv72b@8Pd]U6h ho9J ^ǫ/D(A=X ;)! ȃEtq 0̮z/Yo.Yti:=hÍ`?)RY2dClt)SQX/@>R;&ƈa4BqU7*)$*'@¨+ЈK*Wͫ~2bjSŕuY*/`TJwh'!T {ϛ!h^UU oZ8CԬf[ ރF%k 4XÚݜ|OZvR=cG[G7hcm V͘QU>6Z:[R@j%BRXoݸT`vѾXm6j!G]1dP둇 -Urs`J"s;VA`Y0h!Qsװ#Ҋ IeBф i!ёO\,ޢ Ә TTPtC[B Z 4BsX?mm?83 AL7ZuCɡ bȋuc(ml+. -a5tGV1- YZec``QT9*ДYn(E7Tkt t{zU G6);0N-`׬jPXS©l(`_a:\;G$$X(F EEt(M*ӫl1n2AFhU%ٕX;F͐jPo]8dܡaS`~"(P/ BHP%T&T+Di?hx"WX)𭥘ƒ=k<]`&S1RPk'€:s a `ɦ= plVz bMp%Z| .0 -:;J` lj:Cu;w Zq*.g-׽( E̾ObKR KC e͒`3d5> Db$A -zp*{M>,d3.} ޿ܡG0uJc6(9Yh>vU)*K;,'a / 0gbDuN M@uwthnPQ Ӫ #w(a[T1@Z*|TiwwG݃[`pO*6<&գ-М6nmVܢR;.H]6 |tg҃d &Sp VRX]ے>L]Xv馀k5c4&Tyg'bQ!wԢ`|J,0H5zAdz(q`J:–n%ѲJysO?~6FրkEeW^<6͏WwַoA7kݿ%}x!h`B?'r{Km™?Vxnn/o?x>߼s*kwts/ߜn//pk0#;y`*n UsIRJ͔I7+& zFd{o($$|(IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$N6 G( Āc& ZY@֬> (m$)& $$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIt@||:͔4QȚ4I 'VJ$ tI $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$)'"v$ ^Xp& (m$)&\_+IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ $I IIH@$$$ЧG#^7/~⣦|ջ}pO]{o7w7B -\ch'M\ZR.3\:ҿM}-x"++kY2ZNW2(+hh"bMCW 7Y hk+F)/*=rә 8tu(@Wǡ!:%L]#* ]}k>v&+hoն]1ZNWґ Uv"bQMCWcgӨ+FkQ9ӡ+3A{k+릡+F (JBW'HW8]yi E㮝:AY^?}{{qnn77w|}gPU;@ݝdms}BŏqGp7ח(ލ6B4*?xʅ9хEypj{}8-͹y޾y}9|r]~ʽyUN3ܼx_ݟPK'+;] #O?/OvMcKUmq%msdJӖzoZ2I}{6~׳mj]}ۓ]$nEwԂӵ,ndUۦh[a 뉟INF$?e HԕyLeZyf)6LfC<,gCjk-[oSncͣUCO,Z1q]{qC{'`7ދrbw3fkhnK?m\h]m5~'mt8>K_1i'Z`ac1 7YmZ}f4NvBOqk'c8y33yvB6vb ] ]Tt۰cf+FQ+tutSv 8i piCnj֭X {ȃteު3`V-}Գqhzc1Gu:tuBW_3>1&+4tS*BW֤:I2ϙNDW  *BW< BW'HWVkf+FMCW ׺Y:Z;]1JNH7f"b~++FVo2#tutpi"`k1 G 729S+wcpfbR*c?⺽ͤ*j7 M3\K4j4(WvLh44R3jmg+FK~t( ] ]EVԸC9g+FQ&ٳ;EJo?]3vb6BW֭^]1J#ѕ~3e-qSzf: 3|8y'u$ʵA}]izZi"bNCW3e%D)tutel)NDWL^OCW 7YѦvJ ҕ%c.Mk+F &+~'QY&vʤ ҕӤn] G^QbsO.Hy _p\Ze&ycqe.)ytdU1踒VJvww7~}N&c@"]zFsM<C::)D$4; k4bm\}Z ,xHtyn3Y&Z;]Sr*8Lt MCW w3njֺ$9wt׎&+j)bf+FVoe3xt/3+Hi1ZZd]}7te9Ly>;]8=w(Ly8O|U3P*6  tQ?V ~n\k%wy=hЏUw 8p6?y:|4YZZduw̋ XYThJh ZɧJq<UU56n'c6$|HԎ !Y}(k 79 3 %s 'rڶ6MSiZ.2sZ6捶?ܸtX#p&Of-ciDm$<>>qN由V Ǖ8|*(vBv\7(Տh&}qfaQ)0.>uN8;Ik)pd$KubTkP)L`Qb5 ]6?%T$_bFECpTx0UxOSz3NaIcm] ?vgpk3rI26zVδOIό%jR4O;2ʏ1JٯFӭ)`~}.3{#]||J!aᖃ3k8-@]_Ϧ;pU!D]gpMn6Tq N\,{4p[ rbQpN̨dNi}ڜMu6t]evTvnd4ӲwQ.RlX5ugZ.M2>p@mܩT>ʏ*oٓA5iŷ{:Nϰ_ϯJ8W9Z#J&+թb^v1ϴ(7wUCRB{Z-^ zVuYf-߾nz۾:ۺ>=ifG_ ,Eۋ]3Hb-%a*c[l5皞;!+l5\\ϩ j.c-L: }R *d+Tm+S@g P`Tpj۪OҰJy"+l'+%\Zcti쪋Ҋ+b} zS\S~c -`AӨ֚1*]"[B21E'DW2᪓rX@R} :Nfժq*H <:B| \P3 W/WA"ڏ;2 a*Udɠ Mz9BsBar+T+uB$\uW %!\ɵ P>B&\uW+Ga'+k$3;P%O`q+l]\\C*WRJsiR)]Vu32NaEMpK 3US(Mdj݅xb74X&Ȝ6I*^M*7FXsFp;G&RlTvDa*H)vSlŅr')c =bjv!s0mQ (Z*F>zLJ[` '+l\\:bPmb@e=ROBRgFN @H+T)Ҳ. dkAg1ʕ PV Ukhz7`R{vDa]ۻ5LRkΘI xB;2BSU6v\4fI\Ym]`o FfU2v\J:+u@StP&3#jWҥdXy;P"]Z}t*]J_|CSRE] ֶ aj*RGV'\={#)25\\/2E.3J&gwMo{M_j8YLn9GhɃ38=gFdKT]kcYٷ ۡ]^n^'eg/7yqY.@$/_Ma7|vݛ ?̶RZSYr'V?+ 9͵oziQvj=`CF8,Ѭe ّ\el8 Nϋl25\oeq-_?|6y6~Mnk`V}{4p[ uݗVoNƍڕ2ry/0b Ɋf?OfhGv-MlfmQ=?QIuaڹ;dB-LX=!揌 娙+fhfI3~2} `(y0jZn s~QNWF&lnW9JtlrRm;kvLC4v7SIힻT>-6Jh_&}wVkU+Jw%y}6p%<M/?|S^`ռ>Ε@lx;'/2%%%%.>A1ess6:>tVz5j?rl&u2cZ{[||K'欚?e9lڔ^wV- )֋%V%Y[^~Z^F;_^儽Rs7ZK&SU:Eպ~Qwv c#ZCjR|8cўQ{NgKI_q_#P ~jcX{UOb]>m<*Rp[ڑQNaY%@t|Ozo&T8p*@0se1̤'FbpsZoҵ3lq!F>N,,L<_\ů?ǘx yLUJBGyݾg/{k~c=SdFc ݼp,eCUbh_?2㗧9y|{#6$QRh7O9f 5[ kvR9g- X~_-խF'Lv1;YfI52rr5\P2_6BfPl\wwzhfmqi6Oc6:[y/sYxbJxبP9^ȱ/uJG-g#m9s?R aVXroFn-oxk9{GaNQIu&. k>x vZc=nK(ÌlmJᨌYK,srdN]IgPɆN 5L6l< @IwK/3 j6$3 #;׸ s&ɜ'EW1db֠iĬaW]z|YFnk_PcLxH vY,lZ!wdN,6|XW}< iM&kǯ1ېUcDUDF:LKLL9JoLlUZC&fm\U_K4ۅ5Lgm>ͪ]Z$pW;u\##g t֤ I5SimKr[sN1Q.WpoWKxbh˗' |x[4X/gEZ(xQH%DI&xaw`Zf +c0F'kΪIY 5ǒ вT*K҄EJ$IrL|橱%AK#?-ii OiuNퟵ8lt޻s1N+ \2t"BR\2J%p,JL3i \<#) fCij$i1U3W34WȖ?ġ9 mmx@v:x}NO_j&j@e,L4wpySMׯy 3H>O_!cj /4gDֻA#r3pX> hd;7rmImf}T*`[70}wN0j45j䟭DNn[aRQ|b5fH]GĂ2G80ٗVzay /uu{ubp}NDw\›_巜 V?/;tּ$~ݐ3-O_0 pStbS֟ݸ:]n׳tSV+e{QTV nme D_fu GuÀ2yay;`x(Y^z,pK-߰ՙ{uȝ1P|}pܗN!Lv]0܋P++^;@d|=?3J~b7; }1>zuBLRv~AT&aŵ~A'6UZPk:"gkθY'* v=vɀNq]CYq ![-M3%8'Sa`b Rcc4"S$\d2ļXp <1˫v t!)b{huMo:3NeP["(旅6}>0(fXܩ;v 5b<[ͨ"8#A^OprRFpjV0S RS]0ӭ-46-xMe, 6 )Ru*YJfT`څWg')[DRCYK$\pͥfbYox6*l>@"<7@d29abBKO-ZF#O=,pU, %p<}/-X)m2~'3^Ewt C%y#h'Gy h. _]}ʧ?&i/( ͍,TkS/i$@UXp@|U DGIDk 䄀P.;PJ\k|9m$c F5(Ou!Bt osɕS4 uPJ2Iw^kiW2\,翞}V3S?ԥ$) iki 2#2pXbRL\,tG#c9 %Sa\MdSa= r饱z8) UHhJ-1eqLp̎GLhIelJm>B婔)GѥáCudƫ8 Õ䧐 'pM `VM2xWb1 ρ~j !imFQL>JVoifмZ2A9$2Q6;g2 [9o| Ȱ3muUecL\JF={)hZP٧ /H!$,EsףUb' f n xl3m2`ikځtyW:y.*LIhU^L '!}SQjm޸V7i*H(ns f[GZR϶v4!ᄔ:bHEaejh,'hN lWIo˟N@}Ap>.e@vlp tе$<{I.΀VB 9kg˰CNy" mx /c`i!/' ay4ˣ5`SLFˆn +3D -%wD4 ^Y^٬>FfgzBvut Pq@f"n8Pnű=hA{*cA](a>3O (xrAȌ7=ZψtÀXlӞ=wG(`\Qplde.W1|ōq_}u( .5px!6lgNZ4΄R@&)Vf)1}s9 Ӿw 5qB ϰxҬdޝ'ni 1S~@ P-~n!%B 7l$gxt@FP5+e& Φtt9 O|Yd(ׇgrYac:z5Wzϛwdd\f>̍NV';*be H Ng(Ms'S }k[Ytk>Vڼz_㵆0xMY%$Ij^ *..; ,͐(K)4*f}mYcMa J,tRaV8=E3F0NJX./B)$$IN&@l|s[oNG4{Vx8bfr`G_ 3{v)UEdq^͖P$5_l̈́jPfɃ:T*KFccOB+ {ze۞(=+{/, ;Sj2_)g揪ZDiRmRb6;Wd&Od^0Z "PP IFXxa shLZ,Re@kEvБ1%ƸǸLbq(ѕbUif;ʼ jYcM&Q ,F胞+J݋"ؽPiMGlX,icP}%eS+{O,-(y{E"j+qTU*%IDkVtI2%__ TRvm*ZoqMNQɏGiEϖ5aGE&,/{Zy-kܻ;=;ux@!'V#.J#)g\S2Ǡ|W7Hba`6~E-u/_m?k) `m2Оk/nYBm~_9|q/:65.9@[q^>>!qlrˎBVɭ^(qh"% .~dLk!`j}l1=UyWO$Y-8ʭ=[xrMz0- mZe4%ƪaR:]} @K4LFfgIp#:A' :m^ ?BMC\ wt!;hǖ5ޢz$<"mJ5³KH˓cѸLS鄃Dv+nPEV>zkSС>B#|\ÁF`* QuBHoګ:xb:lR.,B}wk`X٭ *.t7HiȻd}s<-mW[oܺgl@ /vܝԜo"bQR4űۧUil\j(ԑ{Gy42t\i˜|W }`bwM%}}5 rJwq.$~[F)aIÕ$xspgdw_[OaRki"1~N]xOa+f`(F%Y|^K%f:+fPK̛}kwG$Z,.hk{P&I,YPD3dwI S% 1q6$@`37B1lX/aK:/dQ1+)NR ÒM?d7!ͶJ 7Ob@ ZLcG9%L y. a1s(M"M0Y$x&F." i%) ].h ͦ]QDR%zR̰EpsĖH t%0p 詯镶]ztz$O!%zǙzpF 1MC];r ]Ru뽍y֕pxeD_! Ws'1H IJLC4CO;'κ8ZL0–SOQratu5lqa FKu@nt:5D/YXRCF^= ;d}yOg D;R@4$!K!08|rPS&j(kG9vv7kl0֪砦]?%I"SYɰ4`J1IX4&J*5hTW3e\mKMgnP}+H]?n%KmC|ku ]gH,(A}LK?~|OlHǛV`? F*>#nZd$ŲÄZȕ-WnQy= 1Kb p$YOʾUASh~)j u>R)E1IȴCZ!czT#AfBŷߖYi2_(]+?R>vq&kO`NMR+n[tKR)8j ttoXzC,;z (ErFP'*+毅aC!˓0ϕa D@l0?Z6I_.aBi\8" aăg0a"R|v&D1L8#9Y}[m؃j캓GE@RUQmҥ Jf A@h{ZD"x8rWe=KZ:ef"`h<6[Ydsήnm:~tMTCqTu=JyMv~:89?W-e: k=;*gZh9p^8WdSqg``m3C'+dОpI!xO<՞o>QpbǶVtVGDZdQc3_p;wH7+:XGW"VؗD D$-:un kMl_viRYIeR3ߪQ˻\[5KƯ蕱(xM~Ai`aP,ӹ屳ɖs @v%6[J5`;L-K$yPKhXJWmUl"Ұ h" 8.|\Ӱ\_a?`/`k㬓|M|f55r66AgIpdsbhްmjuoQAװFrꄕۏINjoAQG>Kj 4K3!O(1ΧԺ^4u59t<=+1l "R2GaIY}CARw`!A9mk3IμsQqtzUħO F}#\5# A¶;Vs( hg ){ҏ].:`)@Gj1-TxUw觻? ElDaxf@B+݀,vZuZ:lq2e?ߺt,δon}#ޢV_D_! л,`eM,1=9_K0̬tGWYLϹs a3x=ZNS$e\d(A@vK6Fu#:YB1%s80iTw,x5:m~QPnޭ&ʗ߶vI95KLpriiD^Zv%ClL __sdlEn޶r˞f]H6q it8Y G"Cvn4N9q?]M+G!ƯȾ0ȊCqۖ$՟`x:28!@:&J5o՟h[ll1@dQEJF` !x Xjx[ш%$' Ya&ut{^N<{T?=DWh)6q.r9E,w IyZ(1bTسdMA= (t~[pD5廉2Q~Tͷo Q4wy:R#fg%g[xe0e~㦶 Me[\p/?q{#HI>MtW7K$KU`$/ _oxoXUaow]pAmOC;kUm =zzDҟ]-KOoQɵKRj } w+)"")}<HL` AI#FR`ҠaTkkHފWVMgZКޅRg#Wiq%odkkr O(0ALa^DW z uv]oqIɩk4}rYΝΚF!uob~\t7=  .ZP#[aR(G }/hΓZU,KcUa~IX2=$Hj^w^r }/[IK֞2x Uy-s;-7p,'z~E'1gKj!S/ْ{{5zmD3\lYR<`XH@`Z-V~R8pB`w˘Xcy2{q3-qɣҧ.lXB.= %֌*Kل}]KŨoI~3O7øEVo7+$N6-Rި@*;pK9( ɺFn9wm~8.a")xUXFO㉧Xv>b@bWA]Ā:5e]CcAI0Q2ҐMF{8`,HW@Qrr00PE Zu`?ľ!sWQ$}@6D>̑Ҷ )CTMZX! ]P%gՍM$X2Aq5n>[INu˥BWφQ.ni/X.,tg=yJ.RE(Ng-@,a^Z ?1/c͋~xn≥ u72}S#b_aZт`?}" (,3uR^^Ej@$N/StnMu@=FPrA+:fwX>{=ϼQ>[@Ay9֬VC)Dy==~|ZHmw"!fH1X":%ϊS5s& 1#:BwHYeݬe=Mƛ;5)Gd C2~"Qt/Lv{MeBP- $ Uv@{槳loYlp;ޙcyJi`JiNǸdJ\OzcO.Gq .5nPwe ѸOl7{FA[.Sh,! poRJb>Z$o*/'F@ ]DSg)DX1n8 LPz3eKDR=o<p۾Շؑ^|{ZEn :IGs:*maf>IԏEQLKh(Fd?F\= iaџE̿ݕ)yH[\~q,͗awZm6y3?gt%݀,-2[ÿjLt2?S~J*Nsc ޹:57'aLDa>x%@x.6 d6YRU xgBaqU_>&@Y`9 wyZ|*2B)(WoGg jR>#r%UE ;Ӫ;*Q2{#(J:(X,f/AXm]Ԗۋæ$*"妺0|CaWv4'K5_rޛneٕ.ԸVZ:dNV7!pn%(V$B$b2>ns[.fͿEٿ/jJ2k\0|ə6Pbfۅ_ MjkB{JbFӔbBF rEg&䱽 yo~z~sƊVD!O'w}|>Q A,[7bj&\Ce?گ|?4fYq7sgT_7EPUxF8#bva|_C1Q, O&i؁gS xOgW oD-} rʻtT@41ft\\ :$BPۗV0ݓ:.ë(Q42(e=pw_4 R޷8 87uy7Z{ 8}z&N׏<,876/j|Cn7$_݀{-u[<,'ek#[U ػ:fFBa do1^_o4AqƠhͤ&xR {8sHK }Z˱&k>Pq2F0(K =Ezt3m[}pQ+^dqN `p842BZJoh9hpK Z=6K0~<|ԡYqj)zy|M {'j1MS\$0*^pM"oG,G}Yq"> v!ꄣH96C %aCE6,7@ +7\d2d aW FnʍдBi,_܋]|h27b9sxU`Fofjm7hć/c8=w Qm]-4RV2Yo0nbI5nB@dh0tl3@ϒƘ֒*iUҟ0g}>{TFp#YBD| -ST.B5)hG"(unPcѶt=)PC:R鄵& -C'U`aXiZ՘ecfz$6 t"o󻨬BxlѦU"]m:no7؀;>Y~o_~NPo_mo`:%o2ܸYf2Mtʴ}0` ̊3rX5yP`ߚt>ӁG;QTEiQW2::ttyxK42"M@ۜ́Y0{0+K[c؈nEdK5'=l,& ]F9ʽpU3[?v/ATȮ i4˗E+͐`<5sPg9d2M%y1 f45]n{ΟX10deX,z[5Q -0yh&V'|]O]fD)!QҋB%Î;N) >XJR.IuB;s`X# )Ľ>jƒݱtX̞\yfVHJI3y!j'$`G%|Yx0i+: ywu2aZ2yx^0񎎺LL)t[eX6fJRDSBvGTMs¸Q(HSਰ衃}<18g 6G⸨hƠfg[% 4ļ_ YQ8MnQ&W9|[ ieID豘rdU5#ޘ5vo!(GKӟ0Cp:}ShTl@0lGpDZjl {pֳk& >xS݌0#,>TPHSkf>ϔmЇE읠&Daq7'c Dt=o]!AqYxNzu긘bw8rQj0* \[^]v l &9H`H5$ }dBQF+7Hig``p\. %4C~ ca _M3XoSyE- G݆12#G,9C%凌_3m܀g@-^²JM@*w)6]'U, 3/W?|\tv@tӨ*^M1";}Ui)\'ZKnx99sRo(Xi@g) ҷ}x[>b297#?Icb<g~X.j[[pِ=1t7To7/%ظwt?dmPAhQ=Zߕbue0M=m̾Fˤw@zy8}>>ɵ>e__mo7>GKGc֡ۈl!wES(kTeC\L7RLү ۻ rv9.F?RV/Uf `1#;47HM&9\.veuN +v\ 9G TpM٣ < £=fD1jJJuޜaCK2R!;.0KY|bPqT²(<.tU c3dt:r9l[4BQk: p!KcPfRE]5ViOFKq>82˜Dk3)bo'AejŕeuQ\cbې/88XU~-.Pw=WDBE ,9yTO,kAf ?QAYN$ȑuczc`IN2u w?ER#+hAfj/ۘ)(^γ%cIFP@?!M>aR/-F ڮO:Z2 Dw.VI Ϯ@,ܼqz;5B5IO7i՘%*Ͳ$sPsqv*ulM׃E,zqZi 1CW $\чf3iG%>tXte-I,|S8sn*ᠼTz\֘$Eg=mLǀg-, ND*qcTt^(ZdiU5;.Ț֟91I#"E.{5Ϟc/);{b z}*7錨IŌ-Jˣ?/qjd%էT`|*g橜 IG z߀.K(wo4-ۋdQy#nߧ7lyp:fXO79 k n8DHY/WU~ $et^8(_n/6qNg7o,_.%__.\k-LɠUJ×Rf Cyl/a|Pэ5F #=<}M%-Nt2LEo?B{ ǂT^޿4yɅ [C-jL܏1˖gX U?C*r@1թVDR-m,TET#qB9A!%YJOl1&&W,ʯ_Ԅ}BR:p21S73 s55S߯}^ydNOjq~[k0,ɾ)gfg-bƏ,S˯}Y3t["*qdG%tFIjxG}BqGZAQCKW`\aqNЖc Nh^+PiN+0ҶL6':҉ޒȦ$m M4~Q.i6)&rU @S*cG&8~L`UjUhFϟyũW8AɡI3N%pAdLH紤 i}ˎhb#q$5(e/ xޜeMۣfPh-bM53';7OotQOUtuelIrgOL(iwU`\Nk-w4`~ILt˩,GI/=L r4`᫘H]޵q$Bl_ a]'؇E6Dd栗ɑz43E&)M]4l#T O̯# O"ɱ5nwr iwznAiknǭgeG3* 5vIl/s"XhGM%fE]rk%u~EijTRfhCŶu(KpPYʞM.L/֬1x`[/Ꞵ1j}$- Fa#PZl>l"iE+Zla*jzv糤|ӣ&uAܒST_a%J\P6v%׬>Ņ?nQ̢Yc_GO?Ѧw܋O_S0f8ff\ykD[טw\w}Lg\J/[Jp׬S3ٞnWEru(crh?SAؿpfU 95[2>W<.g)m\I6,ݵwRpͶ[sX~Ey /ӻ,>uj- vj3.;*[SAo-4ة]!kܯ4HE+YcJCMcr WtQ_xH8JyZzﱻ){Wa1 ʽxi0gwvtޭe\@7F1iX]C\U^!DžRZIbrf'Rt,TC֮/?qM;U0* WvR$9ܼd׌Zb_g xfO dn˃C%vQcZو$~X di̍rOq0J,M>)g`xq`E$ Ɉb*&J Tfg;(?fHAP&G _i?H`|:lv6{;yzZ}8~ZUQ?(ŊHG᠄7}[j-Oqx; 11zl k9OG8Y]\ ~輅юsӈc ^#ˮA6ꉬ0$ +vQ.0»&vdž=@#'CuwD5 Yu1WDᮍ]qp({i;D4]dܑۢ4}fE=DN"!4Zb]! !(.}h!Z(E46\ F<&(qJIyO`^ǘ(Ut_E7Ϯ[`//XMyzqd>La!3kk[ZJ|'k݄ F 6Vuep]4Lw_1Nl\Hn ւZ=WC`F^D)EFf!B'=r@v׸k{?rBclUHu,sdm=c͍­ԵuiU6>~ ǝ_"]9u^&H\1M`J'6?v{rMbZ8% .ۊ= w}f貢!@5=smM;&pǃwk߿Ma8Nl#eO")L *~ie!gZUTj)Ģ;Aʑ8]Bݚ_4Ơ㫋6b ̐NʽᆫoL.M\nszˋ>DHO8^ @.9{*+aV78aF~Kܚ F)wk%֞_\-q(c֗of=(.slN$.xXwOu(a_&W-%a(egis\ڊլ$t7xm;&Ys& l P}}Rn4֢m(UseCs[V{0HJT31.^)w6jV~m֬1tܬcbZ}ly@I5^k@o+y0+o ud8['VT<^vScxMB PGG-7S<^l ՈH'H䴔 }m2m@B@׮[r ׮1_v VZkڵN)9|_{WJ*{c@myk!/U9ێ^-ՠ ^0w|ňtA5kE\}xLCsԕ4Yc'(I >x]GMEM.kx]/5aCeEC@a<@oi*,@(ڬdz@_ v 3Qs(E>.*x={?:w1ja\4hUf]ly [60LT"jR $aSXZ Jsttyq\E$%k<cEڭx;o _u1:_x|O<=_ @ilb)չz`|Hf^h?&Dt;yxKXR[SL1of7l%"j)C߹!5ةFE oRX3ϐMK5+O/:0EZsAim\b(` J@x$4R++Xor}ۦ*1 OȻiuL2.m -ӊ!O}l׬c5Ǭ2@BS}.< Y}ˮYc'%Oշwf'Mm$ c,GP(7>$C5n/Hn)ќ{#Bڝ{䏯Ӽv\:/C_6HJvdwt|ɓKw08h3 Rz0OtT7>ԕoK! 9Nj:Gu&҅tM-7b{v_ҿ}'nS!8yZtgͳ dNzʝ_S!m+#&,(٧n`N _/%zBrofX'Xaޠim* QV1т vNW;L//"lu_ĩ<=Л} ^glg aR0Q ΄C3i8CcJK<2۝zKǣx{1o--djƧoK@=IKBNF_qWHU1:lJ63笍k$ed6p6H*x_b$`_y7-yRir͎3?.via_ܦW0kr%/GzFi@.XO??.MAg_~p*YJi> =r.`e>?l0|.,|%2PXВ/HIS,gVB,Š#Wǃ?ZV{uAᆈV&IMB4+d5, 3@/Ema%d^K(H#eZJ9㯸)`J)r@ljd JwG2☻ (; G P&G@,0HNVԈ$[[d[XThEKPD'9Ut<PR8 MdDCDG idMN#[T?E4,q @:ILSG<FA :pe#fr *BWp 񓗊0ѡȯ887Go"h@rdE#NcD*Jw9V8(L8iLP J"<;mAARxd\LPN]qF_USK(JhjYZ§`ՆڤC澂PPPPG+QPm6+@}ԼI$AW3Fr,YPXی*5Erj s}+YĄsC7hml [G9XN!ĈH Y10G`RBT4H:AKa|I+}͈*0DK>â%>ᒤ;P[8%֕8ZPST&]RoUB7Iz4]-lYi yTb$dW,u mCfF'':K +-74 {feBԊLjӡ~ڊnUFᐉ 11()IHsT娇2i "M ъ9k;I\OW.O3J. EUI`W r[ H@y%Urf И-a5Xp9CY󃉇A%FP`8r -08/ 9c'ؙ̙_EYj9[N| cnUG֥`bJU8JQTlMΔ9 U^3Bv$<^))1sxs#2<=f4f.TM+A_sx|z4jpר~Fʫ? Lz/PWWZ`2?Ch;k.=qi$-\ZnX.&fN/ʲQ:,S#HM9+@ LhR`ԛ[2d婓˾}zmtƱqL`o de&+w` <>raeʢ )L8F<0k3{ِ<'B)o )7{0UkfNs9NG0aķ ^MoIfR<xfOA0F6{43LrP;CGt:53*nsľc{Z%3{iN9ԡ$h:bnKy`{*\evw@B}څލ#AŞHhGF =NdG;>=%!;#f6lO:Gfdϒs Fv0y욙* ia\fv[* 8ӥnż$hu^qc6O, 9svw>IycvJq- M3U9Cfh7fyبRgVU:F=)#Nb[iBBUL>0 rس6K!~ eMX>-:17˺wW%9Ĝ!e6:̀ uUfrs51de&c"RG c:YiԕȊE`d9]gO) lvc(stYJ[()FI͒mZgkEJd &qnUAvjݑb\2с$h-7"U)7{鎺;%`.f&;xwyˇGywt9;G?jz&̇XgjA-2j]3E>bp@BAѫQ[pLȖRZߓ轟QHVuc+rq$Bϋ~X1ꥢ s]-(]HTj/ZAAjbꉆQ$k16]=ލt&z a{y+nų_]t١c[K|WXy+F̛q/q*L\9߸r;u6OOg_3vk7~ٷ2ف2b)Rƪ&ĢNBx&q2:|quNC2t:zԡp|Sohl^Dem&\Vao,*{i6 }#4ck* $M/ Ԛկh˦NwSo4NUh1`bW@N](ƫ:KP0 R i7;30t;vUQ,O8ZPi`@TR'*C3)=TAҠY`$,D[eE-SdHi0 g+Hf SloƗ* 첇̀Xn,Iu>X3(I)B9b-Rsh9l_ 8譟3( 1&vD:ӳ{|)ꟇONַf}T [C6ngB)/l) ՜=2a`M KAk}9|>u)G)+;0ꗤo 0) YuɛZɽqMBޝ*s:obCifD!4&Ӥ/ S-pQ ̓/:Ã6 &m A]dUv9&4=KJb!FokE6P-^[  QdIe#-,k?!@IQT"j[(6]jMBTɔ#?Gtp/0~Qh&/>X a8>u]A&â4Z*2(f vb@Gp 1zo5ϦUɢj *sMR{Z%LM`-Ͻ6:bk%&5 F@yys^5\n+Rqmnk ~r\ư.(ORޘ“v=vܿ(hg 1c >VK+lqoW!=<-K_^JܚsY˧eg#{n.}:p ;Z[G֮Օ߿xwa1oW$_eR~QlfȈ'i$;Eu-3z+M =JMOMت`V] vyQ0bbh v%DZ )LQ0OcfH2{s5h*yYD`V>7EB^#/U Vr.STH'e^Kr3=sϔRf(md{~]g; !`v#ƨQ6A<=&Nw6PGh7AMy ?&=X6}``O"wk j/Bd+RZƴOg*}!/<Fc>ի_Ϡc~缮|PhB^7;ғr_[ǙCO~P3#q= 6Y܂j1/js̫53& =pRZ{B-ڞ5EqSo4fԛ[Ń tr4_OQpH(xuUIE>-0 Rk]U`t:VSS<e~ƅز+"B֬O)UhO1|%sge%\_8>trϿSZOM񡁁/ɧ@j/S`?2i(an6Xo;+4 sisS{Z~>o[`>dezˍӐ1 +9v'lXeS zd{bvJ=O '3ίF9l\+E ɂ]N*ͽe)6 L 2!kRJ$khzNe3OeҴΒ83 &+1]pɗ>9]^LzZ}^s|SZ?l۔^q^,;۔v+3$3:񝓓÷>{w5DǺ>XrƚQC]KrDz]d'jJl.}jEwi5G.u-ɱ 'iW0'v:W9'rr!tbuT?sãiL~pu͜ʈ~5T Ccߠ)o{1$ִAWYRFa z0*RdXZʩB}1͊‰nT C GdEh\zwW?^'^)^{(/N/'÷'g^lyor0߇+ee=ƍZm/]_g6Yy_q&W&`z}Q~#_O޼>ܼ]+fk|m5f<ԘPcCxLq;Gpy̭ !`m0 X%FN@-Rap-kld[{ /bFSIB10tݱ4ǹE"jY`:?0m6[sA.Jo*I"%ɥ)Rb.^,RL^wgzpHɦ1rTbA2}@wս:ò}SxnaT{whmJ!U^$0W|x'J^!kjצ-#H1w3FA @ 9BEޏmG;·B4S*1LRO:vzMT+<$ZQ{%b xǹYXCGቯ7HiWwؽ3LW{g sAFZ/QǤc TȨX#+aF޴S0j"!;/5DXtRDIzS)gE(&zHa C I~A pEM[XALS0He"Kj2F9CHCO<̜Lͳ0j#~9!SEw87gEZN W.۠ jX^K򐽱2RdׂbɌ~`fs '3fB~! ^-H"dj}l sfHH woWzЫF^5:Sv£羵蠥F-Ak s{NV tľu]5uWjl-=^J꤂'8I~RaRH? 6( 9Ir,%ED%6~ZI2 iGYvt9P\lzͦA֒l}#݆ 0vKՏ?M4b] @U#+;%٤aMlMi YOw-t`X]]#^ NfA{pi@bbT6{紹!v&Ɍւj6^Z6 ct҆CA׬%jEYYnRuǠq]HIoWc)R1&09r%'RNn1nNtXtyȎYȷDdk>`l}ƙWQGmKoWk4(͋V7 ?>c~q3uAųP@8e-(r@ԇjd Vٮ_1} Όa3Ӕ]Z__.̿<3Mq7zW>Vhwo2=`ֹQ+٢IET*ƄեL qzXaf`jm k_k+^;2~U*ޥ (ݑBE*ժhkQcSk/xއ} ާS{Z[_ZvjBf&޵(lHuq٩o}sݲS0GGdEyn4Dva٨lԾ>잆1ewڻVf6h#[N~E}='ealy$HEM}vEJ; wT>@%3#yGvvocC[w.0D9I;O<. ~vا"0vmv+s@Ea,G@miH;mp"Oijos ]'ڲE- ';:sK_W{ _l~h5V7rNYosPQ ꣟s%D|W\)?8:;ܮ}{:ZŸ憗.Z}[s1zq[2/OEc8d逻8 v{e9x? -Ά~"/~N:By5T<㟟N}FF3E}Si70fҵH*fn;,f苴BЋ.DjKt^vq;>,S9,vOE9mH/gf6h.EDmހYHE76.+{iG3vOYv8BB;iL?ŒH;A- /"gfbp^,n7~'<ϕu{RYɦSjC2ZsVҊQhM3ət Y3L֗$%P0R{͑s}\3]etF4-3߻|0F# O,W$LE>/ h|8 A_Le2Ԫ~>Ef2)0lW׿4%R\p\J13- E/Q7$*D2ҤeNU'?yBE\&: np^f.ÃGHD!ÇG!T"N(א_t]"pXMeɤ s4)ƌ΂`Pi9匤CZndqYɽ9y;PWwU B1(8n<bd*VmD&іUCufg54Rg[5T),PMp*S f_roX"oX eC e*w7[}(w=h|͟F 8V2 nqO:=[KWc7JQ|De nf֔zG_Iaiu=df!MJԑ@nl9p!gjJvGSNew 9.EُoM$k@BFQ@S#oZ4u9q/>T\&4i< _ڊ_ XI1dﳷx=J,{@Yk6:%L~Wki箟ٙxn&qmE]C&{&=411Hx.GT QLD"zX=4--ENU'jhj%T|1ƧqfIއ¨RrgK\k!/:\"4*єY ~0rfzs:9)QOY%m uIRђYm]ﭣT22!iBsH_1{,OFh_}-w[u4ՎWhʛuNA]A޳159j%JcH}ۊ]z lX}#|V@ ŽV8 J^T Ш*r%b2E-4i=kDMCv<9a ԨRs}PQiis !!s8Qj˾z 3_ 㻿=3YZ>[@n_vb|F(QC<5喒=  uO3\>J_ .eL-F1 nӟe)9e Bh@xwNb=Vt>YD~7 uMŹ'Z/Km#ˡ :aȑ#?w:I<D l${AEJ>n\c!nLt۠=4*}&q\>2ߢL{9_o'}^s9$n3=3>n(DlVR6׹|ʭk.VJoX=8D}3 )ai8TVu[W 9U8!䂓5[4O(Ls|3[_o#ITKe6'˼uV8 HwLY9C ffGҳ-DDSa-%w0-@HL+l SSi v,h†67 ;(@}Jˊ՛zs7gcʂ&[ƒ Ibb Ig[ਓTzKșgj t1G1au|_wt՝sE>""Ց՝9O+nx1;Ns.l-+nXq~׻;yío?ם j~/u8'ϲl\~ f6㿯W\ߏ?xyp~ޥv~*GmCorϝ{W%/x_{g??~=>mO__-a~pW_]">9?EW}v2;8竝u;3Ndk~N!;? 5f9a$ZA#y~J)<2f~5ZƼCO٭5{{iw%C"mP" մ#Tl5i,L;&9Gw8SV3ϡ W )oRM֖RJ||rFDRHcZa(>F@cj {tZQZF#eBRҹJ.0> <YF`b+Ts42m<,R4LCiCL,,9<1ÄU1P* [ELyk2?ǥǛwk`S»gKkC?K~ ]1ҪLk ?RF()M\ ('ō0tȏeT2mu!V$&J m ]s? mjp;)׿mǔ[8G B-6cpS!1r 9g2趠Vb,4%4g7kQvZ}ݰ]X&46u-T 8a6hQ*]@֫;Xr2d'ZƲA;Wnlc?_tqq$[d-7fU.Vi FQB\]O|>4uQ8Jl'-=Dw #`䡏J#$s)X1N&X Z(::T69K$K{9m,08ղ@8,XJf@5jOں?n!IË|gV[rdcԺ];b;+T)h1BmUud6M FZ H9imyez^u'kN_t,m?g؇vQ7;!!gȁxNn:%Gn[A}̴B:v(S#ƹD L7a&fMl|4b4#KU8sZ0RQnW3%Va.I>D)WS{c%PSvOya:/)w:٧g6q^[3?vW-Vf|t˅ 8c `5@{5 v+SotL@oػ5MsyjȡK.7ȴ4_T6fHNur&4ҙ}s5s|cW4'Ƃcy{<&H=3d Bk3G1&yRPZ}̘fDPO^KV(}@ɦTU Sh7K.&Fl@"f+]DAƠl6n8.8`u\|yL2 PFmV귘ouV+@x$;ގxNa>}+@/Y7LslCm$K{bΥXgJ3B NN%L(%utWF܊/^(+^ؼ,淽4o Q℺:WG)0ju)p8u_ѩۏ$= Rk8oEsMTw3Q5Rg|W>bybOA02=9斐3L jT4D]+MǏٞ|4wҳ>*w\]>sG `}=RcW,/}Js9)Ov~%3ƻW%F?q}gs^wOm/ֹ[kbshD{_f3ՠR5ʩ$2!(a*Hx QKsޔ|6@@cUtgADDDv7SZ^ WV.0 kstQQq)OQ!뢚&gKtόrɘDFHیD(ʂRZ99!$;h!z쩄!u=ϟ/RUXF4Dnm=O)(#4tJ>Ixн()$YjwlRRzκ9U\bs)Ÿ 0&SKL9[ :`P}OhN> GL5Њe ?]kB0AUcª!xqEn?nЍP,թQJwYip~2[aF,e4%1w)EcNQ')wɵ,dCuۥڹiV4vyaTBypĜKfuRAH(̣5O%:K _'\Ĩkj+?kf2[;c5k cE8koA(hU0V&5 *!Q9uƊ0JΉΓV:OrJG5pPƺe݇Jys'-?K? -#tW܁G7|ºḛu]ȭGѷۻjMgN\b"Fq?MLO2-3+&[ߴb?Z8k*cm;ˎ@$jJAMv?xu =_hu|x׺OoEկݾhWxz966f>ӭz폯:ȴa. %vPKl+8 ]6J8~wn:5%]W1M8f_;c4:qؒ$Jatl}^9{KsJFbZ,L` ϓE@p[&\QF  1sBҧsy?[__)y7<g.V<$2\kکΧ[$t%`JAr 4&'yc؉.1Ł .x@)@([/'q}Ak`'ucqZ쵤&8bTʄa mj0J ZCUp:RQ( z@iRR\|T.R.A`QP2rC'k%4Z.SGej+EK%giV@2H24.- @IqY e]+Z\ţ'_+t6gs& ]L"p­'uΕfo~)ɬj#4nwrgFQCd<+h'H/H^&"-ΣZzjE FBtm%cwnn5sr"l676&UȳDCM5#a9me]FQ֝Zݑ4馩#Cj"ZӵQo /+ۿnV=%tsk`W +l&nٹ~n87/[ow*ِ M!\/1?=")5S"+B{uݕBvw hК0}[S&hzd(m u 7~HYg"!H8ODfs3ZG]­s6;to#\H۪CkS=%Xa Ϊ I4_K폇 4:U Q扎˙Lss+pjACHnFy&h\.Trhęŕ $2)%Q5?E!]>)9yCJWxCJgSP>p[a((!S:DZyA4"J4OC@}u%6@$Ow#Ώ.a4DCzޒZovKxhƎf?oJf?xKGDTC  ݔ,I:=sf5e̢4 簈d @Iq]hQJsˡ=)E_i6c?NަKٳ|R$O}(N$<#S9$s4%BS{7vW\?[bCMsnGP,E39U0EcSCȫykp^*SY MV -N"j|wn|hxqwpw/ԍYH;j_|nr1^0N>Ym{nQ*LRw[jaxh0yB} gZ95"^eE$8W Ch}uhpTTTaAJ;gTpN7jNNP✨6MPv[EPgF$@ RO .Pժr48jG$,xp&*bz@.P*6WI/PӤv,˝C&9!wP"Pu+n%hq6zT^Z5bO^d6cRcBEtnug&ņqJ2jil0L Dbp2i" iS`/D=]s ! SHL-?`jRk-I4aʗfV0J-smB2pƑcZG"NT[DYA)&I K PA{P 03+Ҍ$35&$N r3*Bxַ_oCtr碾Ini>k'WtKLoQ/bW?N/h7fi>!jp4 C}(irӋG&~~"_+-G|*Sr(IEG79c=$`_3ܣʝ_-[u^sh*4ȔK4kU2# pju@.9M@0XMD'G,3*6I9 N(՚к!Z~~J9QA#\)BA/+:Ϭ$qvۍhe˅za%)H9GYĈ1qZ&Von2TJFBtmh x[DXx_Kmu*O{dr˝'~QtF=%C+鹕Dio+ur;G)\XH$KScmjиG ^$)LC םԵyCpI`nYV siO8fIf%w}6sI8=p~S IMZz08 I8]& 4ISp*FE꜋n܇X|b[]~ǽ OӤ+\J*&fdmP~;Gq\ȤhVj꫟\W:(f} <|`Dw_q'o`+/f8!)eY ig֒.3iQ̭4QvZ.@1D+z?WrF sݾn7'$ԣzPa0XܟpcLr;X(JE#XMPq\!Um W<]RQ|ks.cN^$@/V唽'^.o\Sr]o?;{JSK}k֠J Nۛb4e&ïy׾֧M Wx fƹypK`hj?kevH4B8т(t#/.jx>PC퐓 -$i*ıT47>Ƨtrq9Ip@cq1MSfBc. )Ԥc,%Y$ͦj/rEi{mWWkʐл__|uqCĉo5!#w~۝N @nK{UoTZ#'E$b1^VJƙhK8A-Rei"\*FJFSeT ,5_ޕq$"ˋّ>v#D1c]!"}9(=â# 9SuuU5U M6U]Tusznm&0Jy(La&+^}~{7h>\]&!z%^p`80W'J `'T= W Dd޼0x< D>XB>ڣ?t˂,h:Tϧ6s!/QGz~ L(>F]y4x]{*enOfts39sI)znV:nּ32)#s22)#slj*$KBPA:J OGs-m]Q]櫚 qǁGzk ;"fp&ZnY6]/5;~uoaoǧ! su!J(Rs}(B\l2ﮧn{ \at,[/ z|U4,!Co ۶8MWڪt kmx= 3`YE-ɴ5쥫05e>ASU~R5)vs3%.P͊W鉲k!z6_иЪjskvܑ Duwtv譋!|0`u 4YrkTK>;G9 +ސ(b=ry#yTۻIkI_qe3gh"{b2oo9ZЂ3{(}AޮU9A7҂~ԿlZp>rkFf: U\+TqeЬ· H,#,bTJGG0 ,jT9-aD*WQY8_a5CwL|,}TgC|f\Hڳa5)(t[Gٍ!"(q[0B8$KfR/XgͯUJ(ұ#~Tי -"hgխ+`s+A3cz:~74nj=.5#3syS O}ɟɴ#Ӄֿ )=%3 øc4=|Jս #%ZrͫOHuN[_}D?nPh; 7uʴV]&9l>[;/F<8;<"McPA0تe/ L{tz_o RYA*7h[ : B`Y (#u`',sHAJݘL~iaGeAom[ (!{FRv#):`a;Ew{w[QaP索WrrArcpVjQBfQQsS'+̎<>:?}-HS a~9z`Dw8')%Hnu`.Gl)Ku"GM} +:ŭ1|UϠP1˟PzVj؜slwUc#Pq )f9C4y~mqxXB_-W +*KZ}܉JũR)0w{Yrt0G}vcTmt]7狠8j^G&*x5DʥTƭ*MA"$8eK(-CQJ5L+S2Hf7>B%pWXk+Y=^_cC; [%-sfNBti(ᔈ{EG,uYxT|YX!T >\QR,V}iX 쩋ww:xSK}W0L-uYPw\{iP]08J+ @E[u[3(ECE`ėV)xFrB%BU1Ảh1i p *S ׎kn,dV}65$nT*ѩ,Bŀ $R+mKZT2pEAC7sP -DR̲N]fkR2WL3f c*TC;/}F +%cDN zd`H2YO䐲WLR4 sv3yBD@&0˔ sUhE4Yi\MfA:vfWO ku䉯%7v^?o&"+.9?}s)j<2m57Wyu O!"*˓9קu@nW-.;f<nj[s3CiäD\,^>z5$-i.OiF7y6)o)P Yޕ9n928$$Z[`6!$ ʡrED% C]r29wA ١5,  2(y1"R$ 2@XS.SbeiH TĔ 줗"Y[1"$.VGSeAt{MYQr2j"j-u1xӑV' /kuh,#u#"dTYYFH!@*lEZ!zka8]6Sp,#`m oMZNWQtϭb0*PCa)8[,*a 41*@nQ:iЊs V.1*ae6}̑B^icAl,v= S6Ex/I`ތn׿<"WE!5$]J3*A&d=Mi_aFC{><I4ZهEX;\b͌2,3@rhGL :cY#X# %uhKh:>,h zak uբΤT!.10*2T)+Gr ω䜈sV뀐9Q;?cs;ѻnI\kڌ] W<]&Z[Hc(xoG`B`%;#j7en?L*gh9Ȝ= UUHJ!j*rEbx .׺:V<:asy3T)72*s-T4ܻgD%XG6 M[^":` Qm`kyOW9Ysz=kyɩN]PWx7Kr!^#`¿,zc077tBnmpS5 c)yuvj˝ -Y3ӿLfm."ޮS;*mIѸW{RI|.J' kю-(נ}7 9?ElAΣk0 z -E˔'%iHs KN a.kmW{Y1,Nm}b[E%nkIbKJ~ȎS/ ml Cb@mG1Oٛ_KS0ʜJ)$l<[X)ZܚEk]Wdip~d 1ڣ5\* (1"lOMļ-BS )"UŪv4&AY,30}"uU;VE!D~^\6'rx NH%#C :Vz#|X I)c(61P #<”{·|CjcOjP] ?U1p=C 2A.Z3ԈgKBR#WuhWEU *N;[vYr=Ɏw,ܑ2q5*QR$0wDGWkp= FPW ~ n) S,7FyH^ h?|$J{[4>bz9$[Ex"1jj-5zT&n!j~L#,V*[chz]tVr|+n%KҔ\Rt]?Y=~V rN\o/w?魺c>_w~zMu$tUqe!P&%mWSksB *pH?zp=Q/)T̵T1pɣ\NYgo'χ5Y܅ث߂jǔ Vޙ?ÙJ^{-,|Q0v/]]O/e~N[SYoƣσ8%,mf 6(K˻4Lpnji,ޘs]e=Xu[luʩ +N͇%4(&*4]W ͥVz}̮MͬtfMMSud4ԟ/Fg ٗb|s`KF|Edv?^~lxc5z=/L]?lTWKno q0d>>EcE!wGl"Hb4ܮ_*(wOisK?B~*mU^SM+,% y`_oɪɔ%ʶR fezrjzفg߿ޤK<$/GWz/ґ-7ّ}/ǫżi6Yϱ!!y_ _sd,+ev}kON|,I'v],;`O~[N:ǚ+?Kv&z{=ZmezgV/dȇwg f:}'e9B+L;}.2xp"a>d2@Y6ᖙW` gRE D ,wG*lu﹊U;L)puuws哺l}Ƕi#g{|9.b׮=s<3t{&6& ~jL-ӽÍX8 h,`_& ZIoOTb39T|ᣛEkx4r-?ƽ";>$/q}33bA d8m63LvϤ{%1а^/i\VkÂ~h_H|Xf{FEz)rG{ߙɍS37{e`[LU"ƾu\\0?[ dx9v}0bMWug`Y~Z\\5H=(4ux\$d9atұyKi*NA<)vQfT7XV` &n-L(,kuL>Q."̯3*FQXA!yG!0$ÒƑ; =ޮ4f;؊ܦnKin 6-%w^Rm^Pһ'+#Zib){cdm8m5::Z[ol&)Eu 7!Qߍ rN'j0S8` 㡉|J! O*0vQ]` F y̻.0 *b~ G(JAIi(`J[ :<[{/9i/SE̓/WTprhrrű퀦.e4%^<Usy~3GrBSO)Bcx,9L[ 1#PSÔEǜ %#FvsǗ#0s˷&*$/7y.P\'91:zÕ"Jvꃑa\9s) E &مKRTk7([U3ĝ%&]%5&"=Z4' @胍kw! u%v)`wͨUTBu%VĪQn%VX2j} Xcf0 A(D0!dO~Urny &e\=e~ A'eK.J )NJgEvV[*e";+"WXdEJc#4T) 7\BHCLHBL<;}0?mcOHfٞ==RkbpȔBT6TݳG܎>}ڎ*%.4 @i+Gmd Wd@pP^) \SRJSyH)asۖ[vHv701 a(HЊj՞H,`o-bhd WZR@i aBGQʀZR%ij0ԵMV/Q?@o4&ƁHΩB* 8 fȾH I cC5R!EԢ)HV 4lHlp$8 "kD QCLqA%BRFa@Y^c4WFrGraLݷХ" xJ%}t~AfPrJ5)^֨tsI*Q}l…I& OЌp x$1A1`b3"M]{|pQ# P>+$X~i:`Un*b۵&/bOEzDYO-/R\a{ EveE޽y׃껯p{YEnyf [>:\>HD zj T\1NLJj@K ԺHU1m!̲q7 D–k[YGGq7sMfU!!hHnc\lڤӝ-vo$<?q1dڻ jk6Ca()5_hW^*yƐP3䈦+t7 y,3?o?oi5F:$E78pqix|pMu`ͷ$ *NErA hN":p2eMgY%Giз.;Z1hY+ DG-p5螙gUlU'};ׄq{a@*i&JG{LmxC!61ȟ9ڧOOF>>}3Byuye4 `%V)-\ѥ1*tFA[*_iKJh|F3ܗl܂3zNːy8 :#[plpjRWiW5|їyonˏw>X/_H)=]$lB2\R<"<$ JLsh(KEZiᤤs<J; (9!r)ٖ\xBV(th)+)`"; !(e 0ÏslJ5e1"UoUi cPV/ve%u< KkW<4YJ57Ȕ2hO27'?c+*'J'$@"`M Oͻw_ꞿg$G gYwK< VwX\":1ױh:&M2@1.OI+qNFϯce4R-PXhHTօ27 <+͇=H=^F, A'W8拕G;U_/?Bn?{>+'zl%omp({?ׂjM3ג}4ˌ4kŦMqPMZ[ Oc_?Gۯa;>6\{ɚ=`9.c1b=> { >-/UmO)my+hqH1Ð~澼~ŻS^oW}@3oKy^.lKp3E)B-0wV,O^5[?GWDED\C50=b.Ef{m^v3yHy*ņkW|ݣ !MtT NU >d~wfwlnPo,TvR2b+\ vpr+2 Yz PW$u;):jT{&R IcɅg4%ů2}5=8{M "xXV|XiL+PڤM[L[`TH5ۀ낐ġ/B~%0;u;aO[c+ 1Ã?ZloF_4j$r$བྷ- FJz٧,*syQa_v:Z4$A ѧ?!~!D#vA堡wԣ *CN==f@!( pz Khpd/v%i;epPӃ^VV;dK])`E>%4k=`<6G{I 0*#a 0I}8 4r3>961-ыNɬwWwfef~ІwYAPO.{P0>,rZM)U l悀$K-J+r!D|07-! \ธT;O_bk )OoQ;W;h}l("nfֵBn䔼bmu x(WA4eMcP6meMnLe9N>;ǜ*jC%u Hs*u:o0s_~yKB݃\+`Tm' z&ȀࢂWU@#b-zYZj&u#0A%x/f'(; 9xM߻ipD*S\'X4V0؆D"HN]dY ¬FIF=ِ/+GEW;6ϳux᠛C .䕽MCpzTݽ]c(fc0q~nw;{&w4ϖAĜ2?qtQPecn0К/F~N&xS˦.ڡvK']omlޱ|kǚq,'<}*=YoD)ε#e| k% ѷxkpxO &q1!ݏiݏisy@B@J^M3 C!Þ㿸qAa7* x  M>}Q @˷T-] J=tҙ},靗wBO^4Pӊ lxZ uQ/Bj *?a@`%)ݍls/#e]nBtL$='xl5!-X]']|<:N Ʀ.\GS\]~)Z~yUތ;h4zI8CX?CxV'>঒Y(.hxK. -/!1I-(bw<&.֚L m=7jVQ%i}Q"?&dSg novt'ZFNRzA)0oTtĬ&{G5? yv_S O:cQ\Bm3hP]%x.+RD\sb\ b@#I).uqf\Bԇ]*b x?kIkٕ:@Qk7z/C"Bƶ ΰPpT <W3(guH dIK+J:Ug Q /n1:9?Trx>2TpI;c4#BvGn]ssXQZT?_..vル<P](nppUcHM`5v ^|I"F\oM'[:fFr1@P?ߺ! #vay\*ՍOOwa?tzyf_eh4R`[HEL$\/EFMUW1i*QJbxbr/5gG(K\Czsu,m.<ŦDìaνem J~J̟65#h,-%X'%K1q0m{m',4RȒT*

v!uM<8q[oyt}vۻ 0CAM~ۥck2(kSS8v=7m;vv^ ?\)'8Si7e 5BàN&M6ҩ:8.h0 .3< I '>8j͹zˈ?wuْHmV@ޫwW-}A>A4?ru,rZ>2vC%|JSjqoЌ'0^Lj^:VM-A5SR8L/Ѻ4">F~.7{Ϗ}:RP=TEEgrJȋ'v)@6F|?.hҸVqJ((NvĎ|xwȈvW]G`1WT2XA)HJ*Z•2 ] Uj+*=`{N 홥Ԧ2v;QqkМmD`l!8C;x=RqNHƑ:mfkfWzXYAiC4E=7+NkVA*deq`U"%3lϲ((oUoOP0;vuKSA#?(?sPqN0*aq63Ei RUyN 9i TT gۑh[2%b=h{rYjYQ Mi;cXFvtQB,dc3964r42 Џ"  |6 t{:f-{gj(RpZLS*8D/&d9?,nT9Qn$ٽ:JРݫXU2UAEC (p]A8HKXDt]YoI+v[Tއ?,g,}nyښ%HylodJEUYUd Q82222">K`Qa]g;|,:K(q8uԕl `  lڎiGJWĊi*c)B1HÔ AAEXyE0wZp(QX=ڇֹ$`S++LHa+ZAB(£u. >59KLP6d"m_\mZŻcAbnTLFXGd!XaOa0 ylN4bD"c1ݗ9$7(&gp`A#TWWڞ^JVJ,,[7.ód͒:&~)࿮j'}lo xխ?jxK ք:(EO18RDAGJd t1GMdAm&-rJe&d-nnf 9=mO8VK8!cⶍJWRc3J=ӡ,ݧJY֡{l(5ųN品tvHOBQ9c4SӋ1f8|I9yLi_mzd7=g֪FD |2u=yiO]M{^RG#-S[n߷O.SI%t͹_39Zz9 0>3n=m6e>/8B+N' w^V-uY)a )uKDKjC0J:1|)r_p^U)OKĮݟԋx00Hh<*cFq@,鼏V&b,cɬC䤡>Uݓ2db9R2ਜ:S!AZxX.@K,CYĄDaX-} -%9.;p h`Y.'drל< (Pj3))OZJۧEO{7[@" `Xb&;߻CE ܋T`Av =E>R[q.ɢ"Rb+4biP?^e1^Wn,)U6NRY»fz\2`aӋ֫$h۶Pv+e,~٬Wa$X.ђ71z=M>lj~C)W7:-+nceo% s=JXl8Z"ɬ#t0RLt>_y+0iFub?4')磻\YG26yWti\"qp/#IK)"=7-zWlL䐱~@t#0 Ed~*F13!|NFX\]f'bO,pN g#4=BOQƍɦ2L6倐ʙci3hV9֤jPFG{X?'XC\y*ֿOw6)`y}b8q|s.qwkƶAϙMɭ{U[,cs6MPYʂWx}Ψַ?Y3Zs2Y>lϑlBm-s|LjF:QqQYuŅ {:qLkp;E>HQ gr YX~',8b)YOPm[6?zn]B>Հ"l+OjV~6 QnaFҐ? A=gb[)rT} n U֭\6UH@1m͗_]U{?D5s>k=ߚ)\M~ 98g3:#:YKN c(:a<=kWkpW+WV}řpkݍfR<JBDcVWrcb#քhw_18)$ߤL`8W7;8M_k>ÈP|vpBKx5L&:V_-nVꯓk{pvថ+0߼ϰ}fZhI\%WocC0כ_o2=XoR~ )|,aj3 N0_E;l?e>hmt_ޖ4"tObnP.mxNǰl/@,RMM%% `b)+)PFR"[a}+ɣP-\g `9G4b-N {@R9M?g?Եnubݱ\& ws''g_}j:"# | xL{5lZsqVfGA)h@(}{bqU:2ϣB"=b43IcS C1Vۊ\bD!9l8eу##E+QH,v́f"uc52#-Fm1~`R,[," J+f;>=gjNaDT*8H%q\ #v  ED$cLG WWGX,HAh/,(1<*0cUgB/g_kO; 'r_L߇ER䧭nZJ`6u$Jr vbsNpTR_$ ;.9򚀕&N(6p(Pr&5gSjNΗ2BᾙrfѮ[\4?s5dL ֲݫX߂Rl =C {v;W %*Ph~ -:{U*J~U=zJ ƦH Ia^+k 3pXlT=RFR]ޑ%XZKh=}B@J=;dcj^IPO{wQGl zbDHXJ"E~'x{`LM_'4#ZCߏ_ 虓V޵ȼQL0ȇuw1z̊=NlB $OħT~(hx}Fha)|c#&w?S,z}c s"}*vn{EBi'"ł:wL.d~mzP/s2v%Ŋ%"ͪXϰ* &Wjٻm$WTu'^!5TnsIe6 gKd+R)@v4 nJ7e6杻?^EII͑çsT릒f[%"ό*JBM̋j^rI !k^3z/ CpnrIu<> x ʇöE(h-?pl_:}PI>^s?`h@I,ߓ X1JRQaS ^)[42'yƔڜ3qX WL d!ZX o:hQ5Tbpf21dեqR1 *bRYc bda-ctahd5YG$ç%EHVTgԄHFDnQ GAI?dp*eR-R#%GLhV;wHH_Myd+|',xd{~H[Fl~H1ABm # ,Vm1"ܻ FM3!$j̈́1knzK}U_D8EXJώm/_BX%U|O0@%BqԯoV5g}<֖ބ?|&EX!䜅sBY9gՐZs?mCڪZy2f.!M|f(Du4o |_g7mQe[7I}dY$!QB8L-.>6M]0S#(1A"ci5 `L},˹kTx~(DXe4+׺ W/@ǩVcΈ }C]h.X*9R/$W-HVh͂@ 140vT|vHcE{\<ģ~s)%e/"i I/>ANdX^-R uGJ!/;l"D񐓠$OzKW{dƔ惷jwus=>}T}ƓMqBٳpϤPrY,<|N֟8{ovV-tvy"u͔ U%{P@QԪ)h*IM"Vff*5$xM@PF5r+KuM ]pB#4(A$]bݬoS^b&/qZ;(9B|q緅RdB#-7F Hߵgam WI_jK7z7tg2zoXz̼bn LES}5fBdlj6o _y)@n0;s@@C @Q z66Н :rN5lJxuRa>KHN~-v kQW8}߸v2aB 'Eܞ-ܷ2^!XjeӆjoymyIk ^LTl;"|6$0 >1L͌15RpMSQ`-?2>%J;1b Jp}Xj|丕j{5 7@K"Br{\>=8rDc}C>MS!=\U.":ɭ_KQm?^oBݚ]6ӲrB:$D{11\q4&vk#T-{ t[lxp*F=<o ZWX{hp(+pD=2JP7~[(:g7d6S‡#U`HW10ELKsb9o(j E #Н-u߈0|oavQtٺܕiKOHpE)$B1Y;:[-_}uIѦ]eښK$#\Za}h݅MD#4YTVS|fWwI=?y/__'I XZXRa\mX&-Itȹ\H* #)$H*-'EϒN!9̊)$f=qJV.GY># Jee4}EVqS!TK3Cp*mzc^EA[ $҃uiP^}(P;mlOaV-`JW9׭hu(#uDL=,Lkw"zjQ436 ؔOu_sJrA[pn>ļƳl۴XXWv;pbe@(ؖmKO1Ac3p+6iwHF2HV_ Jy7ApGE;+$UIXbƒUfc tcrPcC{͵z*w9?2)Ɔ/ՙn(Du 5mTV#.RF+ݠWv!"y5jt 1Ruv M:cnyLE%%+Ndg.,䅛hMqĨK4FIuXa6L!;֢ s<'DyCE*u݄Ý'0ބ?=C(),,̪?N 23ɤ\z#!j#3cp(dy1T`CfBj79H@ Ð ^-)Egeen_䗷6tJ<ܾg1A=HEveJ9D8n_Q g"\ujxGhOeJ;J(J;1bKj622Ze"@I@R MnkPcg`m6@+<*/-iӃX?* Re67ެ5l3cDJj0<ᦸW9>DrSkJt8FWyF[ޔ ƻ3hFO6Έ]3X4W萬KuK ų; 0`^Qѵ=k"Jue\Fq7B1?zO?&9xI+b:(U<$d1+?)]4@ n RJB3ˣq >2{P~a/S#>}]}oKVMHOJr%}}jKV-[ڛj;3IR).-aW{C.SPv0dEN9"} u#=Oe̔~lf-STh#a*%SŬ*M Z7fKc/<Z 6`&U."/G%f |)yߘU%8^(j 8!6*T nbǰ;T_u UHCcPʮA9Λ!b ^-+uXQ5iaS`H)~ّ.>^&&rlwdKN;Ac}< rx$ ϕm?^oSr1–+PqL_b>/H^7ȲgrXHj$n&ref0M!ĤN5|9ۂ?,(}: c!b= BR1_;clbj2QqKjBQkɬO6"bMSxcR^@$ƅF`z`A 711iχs??p? a0˥YȄp XB A\1rcRJ5 j~Hl'Jٛ)-xF\J@q[qd P(wyP(TꂦeQp|im9@^vŤ P+ ڸx@X8znvyTq|YˇYEU'}SQn~IDDQY&,Ξ)Bc|ڌӤ&uhXT`jݽG?B*һXxZ*`H)y!ʺk@ X!m3b^_/` 9 4j( 9Mύj0P0_UёBJQ`Y_Ȋ+eIFk$Uu;h - v F6| HV> >#4d#)V,KScxk IRF uH`M\3O` U"4U0 [>C9e^IlKƛ]`P3|@U~~%zg~t T,&π_*½KZa-F77^ s['G1~q+Z QF^2m.Ԛ[ـkȵX셣{oq =z/8K~iFO5V zF~086L4xvǓ_ygf?] w4G_?Y7:6i1>W_{/_+7c?qI3oג3⧫Om/zK]Ѻl`|? ߃=+;+ |09N`#R,)9M- T* Kfj%1 ' $>P"֝`W+IV_Mf`UZ?+Z <ͤBLC^AM-* #|;3ʍ>{(_c(_})peT~ϰfwo;m^6n`{ fD,?rP9+tSaSY7Ogk6h UJ $btcy- |qY,V@x !59cFɴP^81OVIBgBaov3;@ffGBiuWbw=q}˷+pe]ȹ((6SW#\G)_ c!:"/,1;17oSK`_~:Z5\-׃`w.|=x7y! coDΞ5(B9XYSl0<S,jrjT1vY/G'S)BcxpfXt|:9͐uVC4TnIvJ (y8*[&[^D1SQ´~5#^nχ6?hef h)j +L %QPm4^' XT2&D=Ե 匕,bߚ};kF=@xԕΧw9q{<=H=Ն|f^a/$aPk;h}Raop};`k^~}9^,;0]v7mbۦAWWg4rI7P6두.12hԸ -eD'v*r{$4&EYhBZ:$䅋hLqs=ƉD˃2;Fv$U7mݢ -kꐐ.Q2E>o']n<(#:cTnaϺ TX\кĎQź]07EYhBZ:$䅋hLzt,dߺ1܉i,ʈNU%Ϸ֬[4e[Elb+H4N_ӬRq+H}vȂ2|Wo{ ه!wm Kw` 5j؜/'/b!r Xxw<;uL8`IVTIPԇ:u:\s۵jfQ[9D= fqcӹgQYV626Ta6M"XFӇk9USS8%d/S89uʩ*QJrj2#GWAFZ *r v;{W8to)NY#X': 5HI5rLYj"TOLFט-Z0zK(Z s\]܅ˁ1e*^~h+SJUd3_UݪY'{Ʋ~m^E>G t9%vq>=xJq7i%w!<j%ާ,S;&saXs o!:)c+|OT浅E߫STZAg>|*. Mqx. lGfy;T1`]jJ1fmU.О"}hBm#-`HzzwFMw~;{gtḪ}BtܣN_fY>HeG:ZG1:ě4;ߏZFZ EYz~TB| 뢩rKhn]4 QHф\xR1f]&QniuSdaAS}dFF]IknuBԤm% 7ɚ1- 7#Eѥze&MMr"ܠVlE>,5pT&5ijHbiI"&^"<5)M. ,TB*Tb.nj$ zn/ZCf.Ouy*A$%\.OB. kM{SeSFD}UzSHLUSfxW^"4)!kF4)2NwokӺѬ .% J__5JC;8YapDgG$C2abl:$""!!}⽝WmqcnɁ*zcD#N6ZӖ v%3ۦ/gZt? ~ncc_޽VԿ'l:{ 4A`򃅕߾Bͬpo!̝ CMC r5ܝ)BT$NnLk؝(tZҪ,XZ>[;]%eúQB 6;+6 wz]ۢY WσpOfFH.b(# o|] /ʊq#*Sw9*9.ޙ̹/W"*݄K&TI~ ݄*`&9kq~ubr`"Eb.jhxB-)VR"1ަBF,r=()uRΡĒTgI{F+}#û8^ ؝$Hf˛i+(|$%Qdj*3H,&*1K3Dx୺SԵrb:f֫֎ʟjM,3$*d"a¥\=ʈ¡"> 4oB2$k|T:y질.NѼѰ:QyfTܝW9Cp5A F w]aΒ1pM pNd% /AX1`jY8$C`k4HQ};ȕs]܃LgI%P ۟㌱@'PJ3$ԊM/}]:%PKrJչ?)aWhI[h< @(1lw6W\eP?'C!`B"~(^A*$Bs ѡG})!W 7l|Rs/R&4cP h&$TZ  Xj_B2P'}*Qc<>!/Ð{Uj`@=Kx sY IMdZ.;\2JX3$C/"Wj GIV\'מ qV @4q.Vj Vr Q,!O&h/B>@bPnI &v.Mç)m.K3 T{ü4 m@3(a0v_OC\Lk&&`~bqMo>i`ߩoxs +K4+x6wHCW-}9/b wO?2#5|df-Mg3TE]kQj&Xj=ZL'F &zFf@"fTp#  fݑL0{Xቚ'c?_i0vr;ޛyXnF~|M@e Y.MX V(Jyŀ2Œ+w% j]C$7[ ?Jt̬KZ#4LJIB,m80I#z>1CA:4f7%4yT#"s7w}I'w1 2l2ɝ"WFeݸX6Ő?Z`ªyt53eY6Ige~Xg je9}>c^ef!uڳ7ܫK}]s$DqD(z>3$zؽ`Ն( 2U rax}k1Ü;2t8T%f `U@+-D吔i֊-YĦei|0ogXeQrC+ۓfi7GQ:e9{\lNR!Ă}Il}WV3h C+soo-XD@,5ԇvup1oHk^$CLTNusd P] h]? ]3%h8N$m![ىeD8_~^u9F`Gk xYN 1B!^ eV ϖ75sѦ|pYu7*PF`r  DӋ jσZJmtWK!v$-AAv3jޯrN7ɲ+iI$ϝo lX+$Q ЅĖjRW&8s} >@Rk2{o2SO6D3]_Pr6qNIZ }Uu/'ݺe 7Blϊ (bºrhZdWTG4Sw4`$П%xlhF!҆ws9X™۞[8y &lu|bos PrKg\²CB{u٧qQCKjKVºÒ" =3۠ VWϪ!'#d_o`Y̴IlsoaB:)Ѥ=TJG+A7ml]*<~H@ӿWa:0IO&*3 {22k70y٫Fq,fkN_XdbN]p/oH\_4] @lp5)W+ˁt^".[8w8/Neɪ3-"6)o3زШ3)^3't8ܿTu|&#$\Ԃ߱{_~Iװؿ![zuVhgD޿o~z$+gxj3kεW9x=\^/֣o29U]O$m%_qSS2Ѵ9 IA$z UV$7y*SSkBd]Rs^4ZLSP/8kG߫]Jm4YYӊޛE>m<4AE:F):աع.ɻy-\A)O%az'{b П<9)'NI4C붓wjNA,%"OuX6Q6Cªk NTҬ}oT:agDW5c W*8ٔCԀ#- OX-}7*tQyj'AHB큞YҲ.۾kyu{(h.&Y?+qVKB=L̆E19VR55N&-v4sg}*a|?Z`&eay[aIVI6Iƕ,( [ [陳͒sϞV5"žTE$HN[L}Ξ7Xy"h0R$PkYoj_Y\ߔ<>_r:P;R^&w͋o|1mqEˬHRK!ϔ W[KYL4SRNLĬg+lwb'hj`᝼;akmW&w7NJ mD> 'vmБ02!9YW!+,U?W9,VQ}%c&1N\RF kH$i Pzl93ےc,ICD7ceYhFpl Dbg`%iyZl8ީg]yE0*(Yf3Q7fxN=r,z(@!2 j-!b]oL O)",91ۤlNCǜN&x<ڙ &M uAHkWh4`|SCEIQs [3^X7F c{.69̔ QL{F6eYgn{~QLi[3mwt;ϚC.Mlb=*sĄK^v#@iom7> ,w}aYRhHy4G)1=hj磚y^,6H+I+!*̕C"慮 (SC wɍ~Їx.z~>lzqD?|//?}݇W_?h*g*;p>k~>8W_|r~Ɍ\\$KWs/Q?-cϫtZ% K8҂-{pSǍ l;_\ _tAx^>哧ٶ=;ӢۙNb?m#ۿ"6PÀ/PnQl7 6IMAZYR%9 II%YȡUWAԈs^sΙ#3auNYq}@j:\?8/׉/ie?R_ӂ"4C:9/0*'9%Oy,e ìUg0,KqKt>m/pqht # K&s=|;/ace_*1;taY IөoJ&./Rd667|Kg-R>jjjjW-vgi34KWwyflflflfayq" &[S_& /{_)zirSpa6*k̓Z7CiE{@g?%?Qs{f4.z;qnm_3g Z|+2+܏G@r/Y PJ#r?'l/%w}y(4KK/ĶJtEKCZ{UR :Ԅbd&(ZJYQBsOeAY~:ʬ/{ѣ vo~Vm9KcJ}TOi Q%qW8Jj/E *<ٿhO z4VRչSXآl|%Oѫ a@a1\APx@**܏5Lp'xuZW!7H{#=] TIq'kw_y[C?6ö $5ej:yѩW*ar-鴐 hh?௿l4lBd r"Vs: 3[7*jO<"][9 .͗a2s擇Wd8ίwoYދR-޴A4œ .BeHCE!_ȅԬ>",4 ]"d)\>>DqyCcD"d"DYmHkP^q>w3a0qlęa," q|0aD%(aQD#~{ŏ>&XǷcU`Sw_P\ n>!>2m]֐L#nkn2\3%no1MK:*)qDž-?q![PQ*nk`"FwBORҴF~"+t UGVGu4gܽVj #(n9+"l@(gÄyɯ\1& k@)uȺ(g+B3F;Lo cN*b@jl?p<hl.^~s.>\e&w"b6=|}{ 4y` ;]A&lVb{ej)d7zm!h;7úP|Ύ-lӵG]n{qYHr<t7o\a**t7]U)jFPstyqSZBbYE[e^M:-랙(J zAoLWu& Nα.Ft^OGY_ب6-9("P+^[<ưRFnGmV1) 4Fr;׬k#!s=Z8l6.ݺ^q|d|ٚ'jinS>L|1}nJC_ml%'`En$[(뱢m S-om8vU:ƀI$(βLьs^4{eon`d3`41(WS%Jj@g¾Id3YZӔ+vIb,>qw@{u!u]ެU[;S[3wX2KM)!`.U963k#1?L{C|`/IatO]JXl4!;#"N9rwT#&~7@R,Ԟ|Yȿm!#EqHȯ;)amލhd@z&v"yPlX=}ݘ'QQ"m@Q (GC,4'Hc) '_{b*#کx5i*Y˰Z0^oc$}"޽^ xi]qLݐW+ȗ-#@޴J?[JmT^./ >q={&F6! Ĉ3b. j10,V8<%JĪwA9ύ6qAe"}Cd`rl"%X)TCTL4 Y3j gd$/l~~o[ETG)$xܤܺͱr.i* -1JOfѽ!Ҽ(e~\Jb)V"pCnפӷCOAN5Oo`gE.fGW}BwF| 4IN<0[vto ;+> \QrJ`.KD1H^0܁ΨI98 1qIRRUTd%K(b"5Jq%0BȂ(hŅ@KԈZg8 %>.vi-'srT"Fcـ䑑.A]1p4Su!rP@ m(H#"qDa8X8r$ Z] [rPL"}MxhWm^tq:kLY ԡϽB~o;D 6.;*0r> nhf6M G"Uxr`UVoCLR!DH|Oۼǡ)CD>)JiPM=}&8`Q؄*P|^]|][O7qjtUymd{^Y-⋟)O1<emhW>0ax$ϔC0[CR}&x|;EnÓ_e/NQwZF1b̶=s1GP2s[vYD8-: BٺV`罨> ]hOMShD>>okZeFX @'-^B (BvUIdqszQGCcW \:uR{( l8 TDh("(TJDcJY1;!EFKKq7)/.џ4WooNt˫Krh#»`˕sqɏk'L*J&Hl0&F("B1 e" (6V<"RXP2ML=Jg3z*I}ete<!yudWZ.N$HJ'+w=jJn0hJn2$5&FJw%f`!JD*UĹUb)E T4xo#J[XKz0"ej3< /f姂v4z~w?]4WRźBV͹진;PR aR0`IFvD( "VʥTJ+Z2o:R6sTJiQ,͈4&*T$& 9X$2HFV2SD_x^Qt<٩ؗ\L&K_>4 Vw(&{qXhpiK HH$XxDʲ0IPbE0\0YA-0, QI"DT҄8Ns֞1c@bGt'1˗SLC&<#AS[ `~$hfƣ3&Mk:($%Ą) pm# `Gg+8JuIDH.aL"ca2aCi~:KDZ7wDD$ly"RF`% jb0HJ6Af )  V #WbA!%-HᒣXhx؆*nO@5j"+JK<4Ƨ )j1'j=aϦ &yx(g2 0WT=s,/lPpkSeʝ6ʟpcbc;*qq$Ϛ`t'6T >y/?aQb.: =+v8F$ƳSYE`툺B+ ^f:/0L.Dݝ}8ElD(WqS:5c Ctޫ1Hh9&DBƮ{Cq<"1- 9hN&\]xQ=JB킽6;6FlSu_{8a+7?\ g/֔{O |lC>=:ofY-v*ohYi𾦖x*:[({oc^.GtιԏqkqHO[%cgqs OlDU%pU3Umx3dHvt'.;]U"^=s,I^Wޔ䌞U;$4WU] sMɬ۩Mgw49NYOgv:B/;_ ! &HƙB_-f%Q!Ğ})H$&$H:<]"dHF# Ipw,Ҝe:1d,GNr‡d4ysZ#mέ's^R#zDyr3ݷu'⚟,oYc$ONbҩÊ>,d{>a^Z߄?'.d¨ JRNNNm#Ĩ˼n,}eU7Af%un5meݖW|̻c/10 3S2*wźDkA9ӝ`tP%Cu]>ݹ^nbhS{,O^4_N 3X!aY|^q5ŊYglF*pݒ悠2APw5RU|sO;]KM&3LQdjw/j7-j y|UՇ^*l^B,Ij8R!!WJ3sI!YY)*itZ*q$/p J{ I<3½ZwF$SgSJBOnM&@;T8qUf|n}⸚oR1t+卦5P@pg$DonS^?d`}~ӛI}xST|."ϟG"(2a6ZcT A"&ъs,&A(ݗZl.;PSNP;i ]W4p2pd4XDC|e85A.Eb8 C(Ej']Rt5SENREĪG㌑!8jU?u}%-<Q TVh#nKaDvP( U&Q MoDMvR> 5!LSۨ%4n5[@~K"* h|HnĜ b TdV+GᗖE&5D.(TVؒg¥zkĖHi;2Qz(vOQ15H@CD(ǭ b5==0g:`ͣOs)1l=^ᤃ}+šfͷU`j%Fey+sF0*> Eh sQf!L CZg4"#F#3 lԊPr:tÆ[ vʹ!Ia8уW !W2"H `7~t5]B4x,G[}PЫr? #᜸OkAE:*y~L͋ފqmQMH Vҳ?E5#!.E2U18hݺ1E`Ji#:kbzm5! m[bDs[r"Z$S_JpzðMq)rt]LqjM6ň&$EH ZĞ+edI .ʹRq=<iD) i֗͘@%f c`1Vښ8.Wc[_/gxֻK4*YTGDH/ݔKDԫ8j]ꭼj63"jBB\D d![Y7 u+A蔮u1^֮m݊͵nMHȁhL}KK VBvǠS[U҄1L/ŎTL#W4s7UX͝, 8G=OQT2p530dm0LDrD0;$JG4 I)4j p]Z$G]JX.G<iu_tMHȁhL[ԡ'Yn[) rDt][сnۺ#kݚ"jĹnݤu+AV\;s wWThv=gr"Z$StحP+#Yk4q,j#r#(m$콴Rzߒf ! Pu'yFlsɘ5 U@O$ j 5=FLCBhKPK\+I Ǹ됧?TvWfV~=5!!.%2E'ɺaXRNXy.wwZhukBB\DdJGMA; ڨp zEHM;w*F4  9p-)δP=+Xr#< k ֹL{Y iYMo7ӣA/ׯNN0:&Xc_)y0߲'|\0/ $\- ^{AnG/3P0sdxv?ՒD(y*opۤGwi` bӍNW7zeڭpy*l`iMg«Gi,zy"if'{Z$]:yj书M#'g_B> M b #7fK缡 5afF~ 74=lX)O7q}8=qW?\ըLWd OnGfFʏE҆w2(F"DTc['yp7oJk}{%`5t7g|X<.BqZl>IOx>Bؾ>+Jgg~?L3O>\]V{BaK%۷!)Z'vcJuȔ$6E7);kASjcJ(,m,i.QV_csj`ΎY0cΑJ{7c vɾCC0[jQvUjQޫ ,h0KCQڍ}J#,OKj`I^MԛŒQXgIʸ/-irӒ&w*@}kI3Q'O=i&T $gEl }LfL>4IDm,?}sr+LP>ypsn .׾hLxNh9@[92 "U=0kIgܾBǿ7SUOfN6c6(H r~~zm@P G5xgM.,^f嚓ۏ$f)U̦okUE=E~'?䏛<]A$1?LOOf'Qfy6f1F1XJ?JkoG[IWÛr ]k~:& 3&IwlI0H@ |>"I]ϦǢ-kvqU0,ln\M@a_EVbVu̷"?![wwE O%gނ"qm`fFWݨ~.ZW7R`;}W%wD5r pwף[ݙjkf>ўIAցv' x4Ib@(b4d-c)kcm4ʛ7й&Z9~Z[-obvoYdbb/6;,FmY%KQD(ʛyp6q%׽(&oS|( TmT;ޯz7(j>:s8c }W_|= T+=XFu9MjX?q+4Ļj8_Lfa/m>,cR>,c l2S&QǼPG0\ypN2!DĀ kfSd8+aKłWjmvTЫѪM$CQ$j8[M|{K3Zrc4SpF|khϒ=K,ڳKepIg6ͩVJ LsQ'-'zt"i.]|oiVZMC'r14Y~)=q՞| }TJ%w[E )>5gB ~iĩE9 9#O>N>Dh;+vTY[oGfREULE45|N4\kUEdTvT98Ѥ:dž4EiڧmYvkY˘AQQI[CIʢSXL21YpMK52L2c)(mtݚn|4fEye^ĭ8]ď-lReb 2!>Oa/ ZSEcC1Z¿[ָMmb-̓E)t@6䮟5SϠn 5[ϘA!2%,"H1)G$hD{F# "!'ݞu@GEb̨47.D5It#V(N8 lp &ᲣΎQ|8*(ޛ`õ(=-.gQSc .@@}6]Gu 5;/_3ZM4h-S\5a:# Ft^E(~Hc&+{pt9eQ}$58%OأR[}Vt~מ,lY3K;(bSAsc" <*,8"W犔Cˡ'0P SmSyQ9{%#m`zABE7J[i9'*,1%5zL qy*#"Ê8cG5|N$BSVU!S\@B=Gjb 5*.ljD5[@BCJqaW3b/Jer4QҌ54=Fs쭊weI|YxE}4Aw`e{WS"9T2 nio@SD=띇?r-HP ?3{7u|-M P|\Y *V5ݻz5c<|Ahy;_TLuhC-fϩ;'a<EAO.,Е$EdʗH[<.]ڦ_dW{d=/ |29g$n;'a|MQ!ύ|%)뜆[? 87wmYBDl4+Ic _"#RqEAtr}(bdf2'>6wrXTdžHg?),9, ݞP oe!Dc:t>)6YO0$GqJ*8H Q,"( $^Uߌ;zwGu$Ŕyg 3BNM#y"RbFHZ+(4q:bGjM7(5QsɃaeH"[$ {F8y75\Ps,`\S&\DGQqSA8"0T¤ɺp|]h|ou hR@ܻB4? "NDd#w ƅJj}LC0B.ElQ>O>Q-ZMF]dy w5s%QKJ׺}4:!עgB\],yGYSb<׵5cnExo`/נW{cmXH9/د<%9x<{9IiSMe%;4)=,NlWm7=4uML DHZO޿{|֬y LGPe0u "2rݿz?0/VFe7yu,Xؙ_B9HpV)&pd拿R,h·T^Sp 149 ?\)B}Dns&ل ;~Hy;O!ǀVt;RA'R)Kiq8㭰&by&ä"rKyhwkQnd|>IVS{ uhk'mX8CMh)9m;7f-Z $3%۩q jy8aq(wmBrߕ Vl,Di)YQPT@`iAiB;D&UP&f2L HQo9ǖVdaHXS1)aT/)ƀZPPmsUU7n/ݻ}W1n"RevH"_@\+-{VƸwg_oQlY0n?IdcjERp!Kei7~m#O?j.K100stX!LCPt JqODqoS 0&'eS$Ox6ߚi6)FVk: KA`Jyvf'J/Gh2tE`ۓк|vww&94_EC꧃a2̔&cZLH\[k5*xR:,jBX(2 ӃբV)W$ӪIUޮZdjj C6v۽Յ.'"EZJ.&Ȇ1[5C!%29i;EIm<m*GAC;8fW$G &9 <L䠭WnHG&*,X(4Ş66Ye\pG [I(w \w,ԀCw?fɱc8 t^a%"܇hTNDT9Dmj|--IIodGӭQ3×& q:Ȼ6~_IN3~g`wWpKld|x0|M?dI&Ѕ៓%U䬃[5#Hy8ۋ R>f3Ger1I 2Sauw' ].xuA-(e5)30gK@BK-E)j2hSIVY Ô $VlԂ<RAAʍˣf5u=[e!ӌ^lCmA nfoG X`YacU iim~l27/N8T)ⴆ?2Ԉai|sf6R)ab|8}VHR,7 PDրKnTE`PmJÛEMgl*60!"͠ðR ˂0hq c2!XE8SA:9ΓW m! wG"/`p)`䰡 qyIHK[Y)oy(O ??sl,ʂ!SThE(ۄO W+5Jk1pi҃f0όA$,8BZŝʗt7F[]N'Ef$Jt7q$$,M r,qTU& zD!4it1@YӠKPK!)7QiBCdTk|{8+ gapBksX8`IрgO^ P'RiUyy><y6OX×ta䶨YrmZo;~;3ؘk2B?M, bUNU=W"h`^rϼpy2783-YLsg;i568H`IuF=[.6NI(&-n)ukR6FcG#8>nyU5,ɾ) [g {㵱SՊk0pmв+yccɠۊcx}UlR[o?"mtqteԌCfn9@az}Hz`84O7^%-z͍p|e#sg0MӐP8Tz0_81|mXc$+U?q|>SR _))tHqcRj$K0.RR#E*_TbU ]HKJO,0hV:~]9g`O *4"ј^ĨD*oD+^fHrMkknFŧMUh~Iv-'*oZgrrʅ0672!)i$RpC,}F` :RHKTK("2ۿ PQE[$B!xRzwH=uEpf-2Q{()lPbE=a-g8{Һi4!xG(vb+a4 d^mV/nVPf؟Xsu3eh)GKcC[׉|AHڇuq!G4~L"91^"^ U\ۛ0JW:|iJJ2TO+H+J,B[Υr; ]4|m7ƪ!DuU7C-forjҨ HЎqR:^T!trֻ{\Uj?TX)]Eev6]-#}Qk Ag1~^cEyӕ坷f^JoD?~6i!v-,MHyfLL9bȸ&xw!ָYM |>,]fU7Ty_}[5hbxbꎫq~^siCے<{Oۛ٪[7Uz>tXv2Oo<=ZeSsn+s0qwu]ЦJA#w>zu7sdn&ju# QU xZ<޷Y>:}O;d]y#w@ݷ U4NUY<:}{p{0Y>"Bטj__>j w%=pqEtA펧 .Pn$ⰼ6ˏM/57mmѸۅ_U^5c";o5 &B{|鬩8}KGZyl߱i D1 Wnr: ρ:+.8W mX|fؾV<h6h^g}}|wѐĴ?S_ώiuݶ߅OHD|3YS7wU6%mαz=0Yob/5e44BVoیA}Vi]dָo=2,sm}KI%e}[؏E4Qy+_\ >5Mah$ 8rю׺Q(0E~f !-{8`8ZЙ(0LYk,(KuhA`ouiApH GeDő~oB%ٲmo Ρ&`'ح-ooV7  _ZTnI0m<6y(>.t{q㷁$7e۳tu:weT!,H[ vwt\wxNw,f6 0”]m֛5 UJ.Lfb$bϬ=5wNU^#Xx`2bPsQgy i{D̑"$'MTEtRiA2y,%&5z& 99Vm9&'Xyye+ +W}WBCdDb2#!99teY&:{CR>#ϖ~h2P|mZ<=aMC)pxT3e7b]o X6ӆ\4\NY(Pa RwE*4e l05&Yq9d[V+)BqOyLjD>B;4Ob{r!NmuqS̔Lu*d$'‰Tܵ&ɕvzʹ}:"s Ktht"(l~Aۜ/(t6]ob9}(kr||x݅^wv~#p={/Q 81X O#b4.SV]G 6.a&ԶCqbݗ};ߌ7}Z`8;l{Ux'5f\e?tQSLAm\͕YR=vz(ˈvc2-2 '2ɭS\c! S,@"hif.yB3_og|c1| }va~}s[KJ8Vu4_<}iRIZ8)H*)l1/([5ϝE -^0<#:jY^ #S2^L錀>sz ^ J[y<(P9ҌPYh`ꕐ:wRZ(,k5y1j! PgX`$EDxQqȘYg az>ND 0Zj0 [e(S\14pDDt6Q 0LjXcMF0HJS ?mn%/u4S`P{Kt0dSmc&>@B[S|E^E@t,gFbsƑSGB;ke"P_ hh sj"M z2cIn@fG p9C]e8"SECfzjZ!{8~I~`& ?._e_0 6Stda{3+1.Loz.~G _3C.9~ V&'fF_҃Rb{ SxBmW +:{)nݮYzS>Fػ]+rwww0C7ac]v;]NJa-֎q>f^`$@dVJNs9bӰ)8v5qeG󵋵8[!zv1lbŌgkNc$ŬI `}v1C^vS-Aչ`-hLTF\af,{4 Slbvv1?cw'.X] =[}v1,gksvyyW`nZ1z0%$ dt7SS oܗ~o |'alQLVN7?ۅcs3]ûn r}ǛK}UZ0VOڽf^Sy^{^M|I6ZXq{ x'~Snh(]w{3/n>.#7]?@-׮fۦ75,*եض0xN)hg^ KǥOSvc^Q!E7O{ VeF(_?մ/>O-khP Nhh6⊎M[ 9s )&{hn vB1":mݺsjܚv %Wn1$E4D0wHj7^*ѡv Š脶qv *鵑[xHECB\Dd.vس}nT6NnS]%9.v ObHșhLvAbPEtB8Eu.8%iϊ-T!v[D-*)ڭ䖂Sٞv O鮨bHșhLMk1ajW zw ĠS#n%E[𡍪~r"$SM#A maJH-XVECB\Dd h7i/lP *8)#üry{n/fD3 Y^3R^{k:`C;p V u&ҵ!9l͕T=˦|SdJ ?ZPr+Sneԙ0^+AR^Ë:C+4•B)g+lE c>5kPI=QgBvP ")0HI Qg`xpuR7u&(9e2b)s&姷o[pT #e+\Z-5g#GvXqr" 76$B!!u(@+"#,2 655Te(`[+L&\ `K A3X3)VLdK3yx÷;!T/`ED0XF(29")ș#NP*kͩt1GY!!}E,h&DPn6>7Fʑ,g1qzW{IQ˕4Z@(*B(^QE)z"g``a`gRQυ4HϔY4j7.WzBYB8qWm}OSޚ?w oNیf뿎k?2vqpf+^OO'LC|sW/x]0&'cdpFb,㭯r\C(B:.)hK X;Qno_ujao~¨pP]ӣ#w0|VG_U?޽RCװX qq⫟<嘀i8AxSwemroξJ.KInbl/.׬$"`hIU{z@8gd8-97'(NYM۫[\zQ],ar3>gl ,1on^Ir6&ɮRp /F]<3ԃ}pC%mk}wKXsOj4x,4-~+W*+Q&\ARLvyc5 u,N nÇcpUlqG> LO/xxcҫyV˻_rry6c/8L)'f:ZOC\-R4L,v?lrxtpɻyyq΄8)6Eos[W f+lGAԩJ`t>1f.)LqX .7^ 7w]XMӄm\,wӏ?RxtWɴ #wC;xw&Mt0}g7 ޸۰g&=+AfBɰQGT2$,Z֟slB0OV{ ,%3 ѯʹyDŽ4krq4+GʄI4*, ˷w l*ua@bY VH6H AF;_J A=Oj&N;LGf hhmSi$P-6Mu>NibGGN)֒R))q焙gfr4ZgMW*fSpiBz 3%DT EUT*D=Է3O'փy4z&sH5`O}PL!`s{]5rOJLC4,T~[ƹl~5?c+0&*ᣍ&pn%L`/TL 0* "=8z͌ءdbHƠ) f. ~\Ԃ̰aEs(3/qJU_j-4 ov8H$uXVr#J<` 1oM0d!eG1@e:M1d מH#Δz;+wl<9-.X&,RH[ŽnZsWjhPZ楯4ץv `FG'zsz|khځ݉?uNJAqRTa5Q*Hbso} *8j×z1_tփ5QPT@ ^@3@ (ѣD*I'0N.( äCq3soooN41/~JsU `>\tyو'-!̞in O_>Ouk 8@2.51mm׸}no]}xfܮ~@?Ol+d8gkڰ&Y?1Jj|M'tc|XܜЪbϩ3$s/}O\xc_C"žw=<1 hfLm<{wkN|@$J_zHܺc75Dˮ!Q@'k с>|Ѓy\G%f+š&goZfZa'v~[6r"o櫻l~w]9惣XL鸃{3N}34ײ֪ͻg"Ox&Bv(%`XlERhP(:)V qMu&92wGDTB*X yÐ`epmXF'Wzo79t Wy=Nc:Մ@N=쉫sZc0\5igYLC9=/GR6gj |N9 e|' ЧwTY $/tU$.-pUև{;\x [Å{gq0Z:\hL5 bGIqQ߹.D^ݠNrÅcBЎ{ٳra8lCs4NeR0a<j OXM` 1 ԕ՘;:sz Қ|Ȑ}! xfRj+될41JYF #Qm,#IP9Fq۳s"EnU30?c <ɉD"1XV6 HS[9K#&<֏/o߼1XLno}`BR-QlVwqSRza%gd(Mi:J #kɽQZr{(d(ig=Hq2|+RI`oCJ`*+=\V|;=nƶLg#-_c*-`!BkO mXDGޟ^7rg;eXHbjd1}gw~xSeܧ*jN:laQ.):yֈx>~_tf_\Av^1`O1IiWRqɔ6D'HʗJk΄j-g"o'i{tp T%ER (Lq r "FcttW>-Bz ҊȯnJ.9֙^H ZW3f.)݃&?ȫq) MZ3CCմfqsAګjcR>j@RVWW-$Zn%RP ڇctځt9hg%g*4 14\Et678SKZۛӣXbhVs*:[FL*콍mM۲5xIu3%Oh Z}$5S.`DJGWx]&9yv5lB'[bx&X%WdQCřbءq:*n=M(s @i#0QH!(KIZPV Ē{e-Dgi6{0 X$g{یAؠ @yd]9rv;V02v$S*2F\J-d9y D$Zh"KYD-)U* LT̄t vbAAjN?.unJ}pV5?MV_5;FF-qpQ;k K#ytG5XAJ&`KOiE/[MX! A(,&I2F˝CQ#66(L; hG4)ۓe"HVCM6 (o+%GK &dAg=IOp5ɖ*:8vaDXSf3q V*a!Ar: pVTb:60?PN"  5fTim;O߃yóP$ȸfer*ZhyΣ8UHNE5%BvtFqjG!MX۝USn4nb:Ll^Fu'bgXLN)*7Ο& W{ 3 76ԁq2j'Z2Gc) iB?cÙ%լrWvjȫ0}6}G5U/nk(־}RS!t2U]}tmLok=op-뒤ymk:Od5c[O>jubV'}|=w-m[ƢzHj#qX40 m]˒$q=CJ2%K1)b53|31s-PO^ ?ݸѕNBL݀_9ZO\fbVZ/$DoQKV6:ڽoOf>`~v/`~&NĦ(mag#U4]em-)*0%uoYA) ,W |w,𔌽Զz 2ѠqY}lk1URDk]'X ]ysI\֓ۨ|aC0Mvq? B` c(PB Z`I\Lv^yuVJEksΌiw. BCln:m4 +6̟5U,FBӣr}}eR"7:>}WjdivƕIHV\,h5pkJ3/j3/0';j cM: 5 .`ågur|OGåPW+. c?*("XX(0V8b̖m.0C7.gr#d:RjrAVe0aΡHkapc BB]T?Qnt6SFiŠ N0 U;[YFYX8 z02CqІXhq&n[QȖfj29䗂/E"s+ҧXݜj:HsANgZ->X\6RǪ1|T0;'S2)!#EXӭ߇Q 4t>ԏisӬ`("w0 鹱Ϳㄓ ce EQH1t$:cfgS; 0bFsdf ȯ|vz|`Rp˸;uBUfg>b  LX >F$zMvnE%(`kxBGbAV+dlf3K<8S=;_Mw`Kt7L2+Y|.ӝ "Ēk=KI B5Uo}f3D%֒/32ZOrZFlnʿ{'♨81 #@\F.J2XRDWc:[?c!Ίz}X'(N8TfjhNUpET.NƭDPcIyFdgba,__ y\wf]wj??~[t5^t+D ;0DAx;v\NZ~Z~] ڽ r 6M4k48 eE8 I$f?<5Kq:v&8^lE͆v;l0@1naنKj@=jmWgc}3J,sQ[|o7Ŝzf_Ed]5[~]5[uͬsxB[XEQ0"ƂYs["P)88)[Q)QSޑ35&Kst .$hmZkemh,dX c$VpIX(FȨ) jRLsʨc=VTj4|.M(x!{G !O#mc41YSj8S(£P",q$mLQD!4J<-ij Z2QqLCHЁ@8Ol}#ƍcv8CP3.x>Γm1$gbDJB[ں׉ eB Y .|l q/]mU4=ܠtJMQA*F>iޓ%eKZx0^tcYJzJZۗ$^i 5|]s^e 7l͵Y&qsb1Â`LI:B:i ^`\d-vz HXgu 1qdBCr1"*È VSB&u"4*BjLL$$<99ѾqhHll)~msK g4uzƺ-8ɯZo>m?{wzZ'gu :}Scv(t`i#`VAx%P9 Ϫ =mB| gtjij]/þep4:¶I'Hܦ.?zkįڃyo0vHa*WQpo5+M?Fi^'9oOPf0~r@ vh{Mƥ6SWCk2!H0!qƂEȩر0RPRbCMQ3)Q(qLpjJ㋳Wo.ޜدSv /]:e5lN~wm;|x۾i>Df#AY7W.~M׿lB>z|#gku0Kj3u%5a.l&k:񓚒Qj%!s 2ouB./ cϺhU  E/?܆WLo> OnZܷp -CgoG nSw}:=> I7A{vt|փVA~o`sbswᢏC->~",1N;HXi.} aF9ZtU/5 g[^+y8fp9/w8P$ɺ} #jiEl,4YóNj{?&}^B 7$stܦ.)Wg7?J/SW`5;%oX,uZsQ]Mq;)q 82#Gq㠲=_n J22ooG*c2;uzcN;v?w;#U|LҰZZzL_ۙK$7c/Y ? X8~L{s_MG=FEk]?t)yb1垟=Vm0e<c/d{ lygOOf_,l1[ё!*zc龰jqč'l}&ac] 7z Ccj֥EP(syLlY:ޔRl=PH8S{h4Ԟ, {3;BDѾ j>ߛLkK}՞j,s.pZ TB9<6i܉])$4 al¥ݠPcnL5cH6['zD6^-Q,ڠC ,r1zhk(mܨ4nQiָj)w!ړ+B'W&vKmv/CfR @IT@;@0SOO%*8 S5Zr5֍p4ExH(2K8 WE!d"i QGG$֧(KdVs73Z̥H}=jm$G<0S"y U E@xdr 0VjA KRpn#Ca c$"X 0WpǚX0=F>sA b(tCbk,:*YPĕx̼)8 17RIfThJLMrieg$DJw)U=Eӏ;њkhp5F]&O믔J?rЦ*Vk\C"(kWX@"Xzl[f|NH,EbP\ZLD;9 0$t0B Dc"Kf\$5W3&8t3 HLfwۜ~z ,n>#*]QDa_RlaB*`!2LCQa  N%24@rhl("2PkM ZRL%i W) C_$"J"%ݲNt`rQF;(XQZqXRDa!^uXs1I k#@kƻ&;]FٔARnjSUT#-/47PA1O9&f!KZuML Ƥ!!o2%0n06RU_an#E?0zp^\0@sܱX!f\ Xwǔ1dڋ>";:(5>PcJ!`Jz?VA"/E CM1m@<6b{wR]Q3x[xy@j݈VH5Q+SƈЪ\, #Q+CťZ<]_L+̟eel4ye`J 82 NL#NGS7n<a~6Nl8&ۿzƮ]/^:v̳봅`y&R! ckI4’@1bXIYA#TKEC, :\PЗ^t2fð?fjGEF0`qOnAGlIg'N?N?2pT9 T}TD4+1D/Mk)k>yq]@S"y,\d<0CN6\a=H|!ڭGO0|ej~ٝ)N[/)Yd9,VD$&*,;C1IӌR*,EXi p ϥ< 1a~*dra7RTTW$ӒlSa2vR8n0pHS!9+ t'BHvT@u1C.|`2?lhN"HSp$VTÒ(W E"EFMmI*zԩ,~;f\{DH%x-sZꙁ(8A5)ףG/bH0-U'pȘ޳mfQgy }5 C|YFcy=OwTX87GZm<qVF6U@W1P֝?Siݎz+hO!MIڵ&hѷK͙5qpz Nw \FQ@P$b P" 8?o>G(Iqrs \+>,$,D* J"pCq\Cǒ5wFffJRg."!mČ|#C {G`mW#+Y@;c*)sn mG'$V(Smo .]Z͂s?9}sZ]Q%߃ eb9,FZ'eW_0 w8'AvfQ$< ^ 6L`?޼p앫~Lӗa!Z~3#QȆؽ5kMPhBc(&֨(RXE\Z, BBICd!?&zٝJOv 4'vnM \5Q1z"r)9q4?|kU:zpԙH&̩9ҙfD;]TXAR{3䕛g'q SwiSh3ެ^3lQIgu i!_7²bT7T~`H^.W.`gpxr,"H-(*jAVAԋXz-swaJ}qn~. ꮺ.xj,%'RobP)EkY *S2 qAZ J R -8 ~ 3V!-'!](e@ K\L"ǂ 뒐;*qbgmޘa?qzzs7io]EĖ6$N\ԥ׆p#ڠDĈZA bB\2}f"1u%,f*"Q3.|67pixm#flJ~77H$K)'@×h8zD:ufU l0 NhGCo `/<phI;<MhÜDUJ|IFqåG{!r:.A/tx]?˷?^<= nm/@,ël,=[^_p$"aݹI{0t*w礁ٶap8#o{?h>>'M,{|~-ӿqX7\?Yp<r)-؇N drkRzݙԴdb|4GSf=}xq맿x24Q& nwvkFAnZ|E7p-^zޝg[o@1 zo'7aӏl;JydƭA?n`?غ Qw+7^gv{CO`ֽ_߼s9`1;؍&v;<:BB Qݷ(Qx-fj`iiAFDf۬MF-b^Oq~L0`t3\Ax|<M^$K/$|+~TOSmen(nS=]?Iy6ZT *҉LwZ2áf{2^"+ :g^gF  bΖ +>2+Aj(P΅fYPcN?ha;HA5ښO,}~TK)93 e,e84lSlS2=Qā<AxAq3 謎r=ٞb% (ƕtkU^4_}Vs_ھ;^9&G?clB@w #|W7GCd.F|to柮mLe^R].kf~[ UK{CyaI sUs lX~2*33]&lMtT3_z!?iR26eWimN~rP BA mM7! =GW|{nFPW| f|uUtYb?An7g v\LO[fZZ3t"!7;} ;>r8>طaA8HS!{apwYGSX|,FlHtc;c'r:cIؼ%)d-{dco^>mho1ߏ-yl%_0rw>vumS~cg? ~vŶZhhMv"u(n'1D.J -AŚS.uA$R{!:ZXLEOK pI'-G2Gş 7fBBg Q\J"3=F*(9~eF˘2f=.=ULL0XieZ͑ZhgM5^9* '|r5tEYA=֒h-0Z=D/ ']Ipe,tY-Thn8+'v!8ΰs`,&L)̵ Y2*Mj$M_(t$z{J5Mٸ#5;@8q2'T"WN L3K6cCi!UHhyxz3;nS =J7f{oG{¸՘SJ:,k qjpxCCsDǹ\\.EBߝΣ &- 8wqcnwU~<ܰ( ˉmVH8 }ZWi0{±ta&Qt924bJ0ҺiވV9L0M:|AQ<ދc67!=BeǏ»H1sJ!'pn~(u `߉z !yAJ+ޤ`>5S@8G#n&|D*{DSwyR!XB"M *@bxr%f>l:lXUN5{Ԑ,[Kvm”$L8b`k8&Pg(آARE\$6BQm'2:8D$y<ݑ!\8nS@99\hSUq\Rdabn}7 ,_߸-t}J ;WsAxQ buAL\ʥn3\BW*4EC/om o?N?1SS =>$KԖF0q PLb ,\>XRJ~PZFreN**-c*,Hqg 'd+2>&Mf-ndg07b탴\a593@UP3qŴvoڱV}fTzl$ c<(+юW9muٵ)/Zy :RFlS(JŌδ`EĄ]g : wbNnӆ`(QV9,isM6;BLc5o|Ht,9tkN-BGmCw^h)iIKq2DŐP^s`Z&Mߔ!2&X 9e4(Oc)õ&HBnՈ*#bPG1}HEԇ+onfW_̅hZ&?\┰(yАj0eQytHq|~k}ߎ/Cʼw28 WZ.v15X_gɩQy>|6^g`f9&r| 6:/e]̎-81;`c.)YQ+eŷL Zْ[b+!!A:3؁c,˽𙓈XVm$?w~:>Y"{,cMQ^5*cRfa(† C4wD.%AHEa$ELqD6fÔ`)XĢ&OǢZj~;]ib 7J4`Vʘ%jƢY-H7WKP6sEJR8 YL ֱ" =cJ1XE4H?"RX@D)g_d~7G*_U8a&F0b[͢ۓW\g#)' r8Eѧe* TL#Xjiѧ$^&Ke잗Nb?4q0O=+=·hs#1`ߟ R*wJs!9QF{"C O0D4\[hhsK5XksbA"[\5y|e>wi9+cdСU{E6R;%TV`Ca6!^G24aqoB.RqéVj5꘱!&2fd`^ƌR_a[ J)S˘Q<# $PacA(r1 $WƬoi!{^ampw[7YO7K?N1 َ1G)}TNӚ0hr"tg,/|~8͕~2k!ւ!ݲU Y#X?+\0-'p|'ޖ<g#i~[:pf|nj~=|~77sh-MD6E {Gq;!zϕVeqhߚډzQ\'ɢ/?9_6=96FzbaW>ojPTnĐy3fQ͙Ӌ_jG:ԮABB؜+߽ћx)!jRjm8B-90Pa}@YQ.u."PX&h8[5/|Vo?^狡[)=/^NثO3烙2^0db dJd4qg8eR]aB /8`Y,yWq_Q"*^0qǭMn&qcՋj<7s뱀oPV.C2* Y,Ă(FQxmo@2p;|Dj@`PcVP džmX!r]K5.\*œ<7zX6+uxeVsRCA+C:^*gR"'BpVOjfpeOT|:OP \AXgrNbZ!B]k^dHBnSSXn O |7OS\{X0 ڇ]$r=,#1y6[Šo>~,bgn 1XlwAYOh@%ϭmA'-밾X,VUkGE G72.(G=9* :^$7=|guuVek)M!|VaR *h[X.X]'j})[An+pA /hUs,[! tXH*ơm30Zݼ&=-jB -0$:MtE+'4B)BNW'Ojئ>#zOS B.j{VHiP=x(c/im)Zmnp+PMKMF ?}6^7qHןS/_}8^2)5{USJJ?}7vik=U0&?"R˫EnW["1*H\REK&̢̈́ %=︻ڸƌ5*U°91C3kS9GTnx:Fh2/H<29$t:0حH{df5$!A>5Git@*h1} 4Qag㖣zF=+!1{tcP(!Whzx C&WǺ$+T{C]j (ֽ;xw~\ U{+i^V +B[#Rnk8 zvҠ@ Ih#dM幩!Vn:M.dm9.<[i5*70 7A'KMy>iZMRI׬ D׊21@~}ͻ{~v.USl5Ǐ:np9/,E}Ab{47Y`H쪷:&{uݟ<Ζ2meZ_<Wnu"~dzϷ+"x &}7GqδH"Mr?w^Ir+"7ek"Ѱ5~]XiϺ}\mf# y*ZS9jM7N!bNR~Xn,:볫~)^J}^^\WNg3^߽T7kO-㨣Va׆\2J8:n^/٨YKH@q%*v 5JQx6xXwDR:Er"︍,íI}g.UH ͮn,Wyqۺ׏b Xy"ht-P'PWUK|xw!6:̘3#!o2fw_zy}pfZ$xNj?cY\yljmr8z^}MbADg[ <&O6>lEA3repp7wLNp8UwC2Mb=:;3s$@1ƿ%HA'H: ?qha1꧳gIauz}Rf :k]vIoQF4Yd&Ç4>=_`W3gV $DU'\:3 REP+PAh˜3ȳ&i7DA9{MqSޜS Do߬SvUh$uHzc=5!( NEIcPkeQ^GEe$#ۡ^ox)"Ë݁7 7?_1߿pHApΐ1tƸ 3\h=GtwE%-"g4/0bS f;"O8sEnD,G2<ىY.OqZWGX$rEX-ljl)FIq)XpO(*F)_gp(:N`r|(APJV[9"LX 65!` $J R5<4< !͋@ch,ںӥkJU /Xp2q1 ycğ c].@틇Z?9M}B[՘W)"#hf&y*}ENY1dR,e($+*7pocVhZk<^8XG @<w[x{tu3Fy2@nNcC=s+n;9xȠ08"ɌxZ7C 'RQEV˸6Q(o(18EИHFևQBf:R6I}g5ۊ&3Q FCݷDq-:>{\ ˘A]-BY!b4Jat@u.>_a^{jN(^w%juXCN-ǃuX^Űc5G A MT@zST wӣpo]J4.YR~>Ֆ$N5=g7]==8I"Fyj(&bǎwt 0HI*`^w9Gt߁zPu18\Vӳܧq8V`2}4zzxrDr2vVONa (ӘvK`j:„Kv 1fPfmNXT+5VS|҃?dtÀ4H 7 +a8&4" e\X8kFC."x·`L՜(,K #,#$ KGbLP @k(d9(Zhxo)DP4S2n/*D˯],Y\s}J2\zk<.CI57:߅Jr&:YBBIFlQ$>Ü =zggWOQpQ@H:1.%(%ttEp@C%.su W"](C9w P+]pyӝy͎9Rq-GVrƹ\0j3_KGCV^pR˫/^?nZO66x]d}uyZEo ?\fRūM{ue%Î{w ~Y t%D|Tƽ֭ͱ|FZUF Vb;xt.u岫Z_>+Nj]}zԃ߉5t@ACa:&Kp6 ,<;)(t'XnLt(H\HQޮCbTOgߧhMct]/_8ʤ(^: /a:YUC[' 㵷Z('~$`Xs|,F3fRDS=J`qATS ضj9%Y}K 2%ʉZv8Wd HqӌeE(I]LN3SΘBRsyH.2G/F핒ogzF_0;k}$LNldYG&"KH[M۔Eb"뫪꣺B}P"4lsWb&}<%\BY`RF`))eȖD~tK yuLc:P*T_%QXjVdF5 fy[0 * U T.F(kx͙BiQƎ!ZXO8cai.R}u)‚n)X+1&eNqknEaZیqO-Ji~Ռ8k78z+= οRՌ"(Xkvcu fNp-F-ࢂ+#Fokcs! u` '$q$(`DCFj{b 0GߗktEP D''S8. +E"+ 8tTbqkW+{ 1c^@1ֈ(.E67 "FW'LsvD m)^i9 yJ֞ A֔ BGe!@r/1Ư.t=)A`6ĿU7z8mH`VjӊД81!=r{fY^n%è0n!ȧn@@ y|K5_Gw[Ep親}|sێyŋjxvG&i~]6M3i Mjz&TLl?e;ņJ?*JswvΎw'g:~;<ɇGٯbW5@ε./NܻL$xOT珕uNQr@߾_[**w~-z\*RT^~xbԺ-R(8Vdձ:o(vW篻ntiy֓?id_f#vvcnr9u{quʹNF^a*Sηc\iMWZoF_2͑w@'`C;Ⴂec23SJ4 fwfxҚk Mf97dݜu`jJQCikwzQ7J"w7A?80Y;Qopv3YF#p F ~OڿhL8d eϤ~_OľK&giio }Фwq_y~rј??s"caNCߟ1O4ոg ff3 GOJ穪!P8k,<[zE3 nᛂt ޛy۩InקMt ^8I]KӈD܍]p nML {As f9 E'& -wW ۖ= ,8PqNpfD*a ]!s 6.|)՘)yUA 1$y*Ư\aZiN? 6/+Y?~{ sgO2_9вq$-' ժ:WEu;ǯOޝqv~|_yTUBeԗ.><,PrB|Y?~u۵Qpnx=JIk]Rձ}9}:Zul_ؾ ڊ,ڳ땢*D7I/E8戕5NO}ﲉ],Bjc5Bz4g͊g'M>dZܐWqGx .av_6jS-0_ FJL>=ʕp7CO!uI/.0{:jEZ_2Ғ=^J!"zkE/-uT!T3ǃUc̶ V2Cͷ6ΰup2TGP(P(z$/KW7V)W5&by/*:d{<8 ]߆vՑ@efjjIVMzDks8фo;Cs2\a緊L)Pѽ7D}&HG@Ȑ)7a@B\,PA "xB37ܛxKD4}Ydy:d"PR0!r"Q$KO"01?!Mt[] 2x&LZ[2gBQZpKq!ڄI݃gTJ4.DžCX]l( TxE*k)ي`9BŒ`yA΍~\VE3RWt*xG8&zKI?poQN.7vnbstcǍc6Y+ىU,};w|3VlꙍD=&ɆoeSMi5+ѭ)6ad!F SL61+2U5͙.>T@$ґ2Ch'R5'_d0frc/Eʽs\'K03XY!>_qȸ묶eBsOE-Ao@ ʒm-u<[0ػK !u_e2dx O2MsLlN?\*q@= t'eU9ۙ閲dSA1N_^O Muw=uS>J>>\J N?.mT(ﵠ\FtI:ҽˁ*8RAC:0kMx$1oB8Ua@Vr˗oŇ5E@/08 P ·"+ױGN:"u#_1TeGzxQʊlj6IG^F$W9t`^6wb{4;D^-7g&#p 相6F-MYx3{;r3HKFN M?Jd$ 3I2htᶒZ^\|nAv?h̿nsO/HЄ $:^?Qڼlj]ʽcḗewA 2\8c0N )#׫ Ik!+K{lbн6y0Yc ݤdcq7qQ]$.sy2p@]YoȖ+dj_"qӃކQ,cM˲[K'$ERTDIl-Ū:KYI%6<OR  )Ek^I,?[`߭1S:+rf~ߛ R'.x|W hjwb`f0lMX*ʀξ_2ǃK} ;|\;߻_R|;SWf a.ƄTs:_, *Rbѹ-&W[`k"OD\B^8DV1F"&qlvw(@nBqUb> h,L|@p `گW QjqصS1 :}0ֽvOo,ϼ qptl>j`>qa>?a&2FhWΌfzЌz3[D>k'.O}{.,JϗY`d=#y-[/dgI5h1~2UiƢ KAkG$&y9n <{_"ocᓵ8:rKI$ázbp*`,-YEVG(pu/_O͒M?h_ zk$D/`L zHDv L64Fo/QYO4f)€IU$ @Ph3ja4.Xj!(eLlɖ5!r1|prF;C']A =ї6mAH؆t>XjyHa >uZ}˲JL{mjPy4jUy绔Ž59*B}$rEvuIT]uFU24u|߭Ze LQ8Ѭ ʝhoTE}l2oT[[RZA)),xK:v#ې倣!\syg6 b*gŪMߌRU6|Ӆa-ɖc_B)мsT^= !:}R ԂL2 >&Ē*&IsR<#)[[Ζ>9TČr- wYeSlMq)nN9D:y  a/NeNI-L.($TqN9>N9>/+6*sCHx:<\) swd``LҀ]joͽ֥/"0?*f@1yxL4Wu%'F;$"z gA?ZOel%5G .*Զ;r~BgDr,h+2:$t25(t?`T1\:mnQx1)*([J=6E,0iĜ"ÔӊaBuOTZI<'RX ӆ0Il"5Z:fT8II~}^w @\Uc-c,sD혛7bd$uNIaq6/D-.H:[q+b rs!U=y."jbDCQqRdƹ#$"U c ֈ.LdcD"RC6zK3!n>kT(XtP~ h!uʔ if HVcCIzJ\ih$MR:a7J M[[oPZ}ܤ #nIP0- 2M)u6<;c3ЗkXNn [ i͙nGYJ IT31c"a߸>Ȕ)R`yapslb1l] i(isr"ӊ>Eެs-xѪ<'=\74`p 'h.ߎp+,c`n:C^*J{#3}<7NQEvVJN1 ׵k80(X_!G]pł[Jp1؅Zoj!H&?a=1ǜS.B|$$QQy<b_)) iL$pNJ M(挬8 _Q}pjǫ_N6m90mEH޿JN>-.:bתL_^y"`hLm``LfL nT2+TJL#.IPH)kJDP4=$ud(~bQ TK7,-J)?^6'1;=8ge˖ "An |xblwFi =}oբp0#-KNlrr[ pV-ZIl(s*\W0dA AZhZ͠QE:Y;)2SV{MǷΌf'*1u/A#}S{J Q9C)ɷw[ܞ YɃ]4*ER׻6c|M-n>vә[{gMgp:<~p{.O3w/PYcMg؂6dϝ)6$ҧ2כY w{_rXENAz.^M>Lnz ̮}ogknW܍zQ4fv{)x353ޅ}{1AIAsȤD%%b@#M\0q^qTsݥⵗ:ϥ0@n&t:Q7OWxr~7~—rshf.~:/_\Og~4uѷ׃. 8OMq._|won^_jFs7X=ŇR|?t߽/]r/~E5`!c_-?Gw>^[^'ts>Cg/??{WƑ ?/YൕxU`iH,;1 hFÙ:U;Y[zp9}=k}FZ{{nߌy^gI OyhEYosj!e~Y4qxݗl݈f7L C,=25} s%g:`,Г,Nhm3ɸM)/ 9[ f$-7cIaGi[z/Ƨw4 fte AV Q~gNmm&i'}{xxօ-we[j)^}6S:A,TzbZ ^l!z"hoo|d?{opp?'އϧom3z)89ԫQw-C} ߞ^woO|>ɋC/^!!X՗ 3UpgdS $A`099ƌC/fQL| /;;9snu~s82dq*\f OF }f2KHW߾HYV4Z^&\ ^e?/z,$u[o,,̸B8Z>Y9/t` g 11UP,%MsE.pk4MZ;7T3:}~~iN.yq Xx%jlwK<Iĸ>!Vb,8Hq߁f%KUO'*:r.F/(xϜ7qуpn|XďW)Vq~UW`,` JP*=Y㭲'  7c*txFr0{åt*(ɤj)$cDːDWx0=౉FJ)EYjFx+= w&41c9k`1" y~NV(W#q\E@%=!Vg5|f\poDpH`:H a > Vp 7 N0iS2 &C(hh` ]R#qSJQ-#"x! ndE  (rlakR'3LN&w*1s+ |KYA鶗8^\iIiSb!%d*},F6zsi!L@ 9e 8Og*28HA̭a5%FR}97`2N5z 4`DD-}^`!$GTQkp}+?'w S*7SөޒzԺz0(4^1lNngvJKǖi}-msӇk%gr@hE,HTHɵHs>MQ Uo)ɻ*77@M8* Sxak0zE%0:H'%ņxe!,jTZ(tC y8 *`T/S,ɻ>QVSi4b>j?uC6]{zqI_3KrB _\\Tٻn\ Eg>LqtK "iq f(!: gkH)m `]#! ?;< ŠFf E0* H"P ~2;qOjٮ-Hm?/2yrXt Go8'f7 Fhtv/ln`p{Jeigۧ`ຣd*W*y&`!0B,n]>;Zw\ I+2`]^6c"0t=tG 1eF`*W>:V^F\*(^m0HrH ʍ[ڨ@BmifՄL/UGAЎvjeO;HudbL[pCpB2B jB2,QKשZj4Hc圽Pj P>L1eGSQLu5SnXJ2D@bJ2{ )ePL^CTJ%us޲FA4L4_ULVQ14 CMbqSHJEȤ5NS6eV l UDLw8ow1n)Lј5'ڕJv\D\A%Ѫ!\de!4"42I{)28rA n;Дe 6 9՛\ZWo,`g//#lτ/Nztk5ŎM2걈4;f4Fjiڛ{ޱfdVG&Jb8iY d)Rw"[p^ $,tu(ށă`2cfX!;B](0k„_ nƽ2S{q' nՄ U{Wϱ'ѯ]kq,_*UqZ+6O[=dbvRK}ig@6' \*kC[^ʦlmٹL_vrP>lj8pimTdyKF#%At3z6WZ\ x;apQ픻Snzr]y;W~˚~4TJB7# Zh"w,/[ڨ @?)+>;_qTnOAns8鉝%92NUϷIU. !)z,Qv;rSR~%pyՌc|~N{vLH`J^s,ٳmy!ѪG e Yg0{IO,@y&Ѕm IrQU#ջ7PxOS F_򡜕g!%Nɍuux䃧A_8F 1U G*ﭧiy R Awdy0iŽ} /I|a]^Bxm 7HN#޻MMf4uD{H1g{ߧ#? ?l@?mւ!TSV:WucVs{ˏLIu hqT}+Ke0mMl3kyH"$BwΗVbiiwfLw5F+ +ĵ=ߕ$X*=u雋v-IDg#[gWousbpk[1e(iqXw¾5hVm'QmQ6wMe[JS.NmCpX$RCعSf5rv#܇Kp}'P y&(4AS`,ٲL } U BxQE+l$%XzHv&N g3(B& {U!#hGhHfUَ1$i$cE Bz'y"BO[ʑUE|G-}T(t:O:ۘ󂅨u1ٛS r/&9DZ7l^"d%Dl#cw X0:2xeul:r7jLTsII~7ʰಈ{|ˣb|udoNäNnY76ck"JJD P<ybi[e R03~[ iP%vR}\)Z9騹A2fQ!`D46 GR5(d_q;]S"Wa0޾QQR09Ptbo8:sWUVf'3~Ʌ?y&j~+*C0KR<j>!X!4!3kN`oAT;D Mi~[eܸf¹jw~/WNw# ݍI6FUycW& ViCp!`Xp9uxo'[`Ud:R?VaPM93 ᶰ b8#M OUZx,`E ~3u P|=&/0H T`n^֩^=|p] Pܘ u )vRbڂZr탺/`9]cD1DA> MuƷT .!/ FXcI6.ly~">vߥ[;;Ւ+2fv|ӏ'<]~:ETV[NJ=pZ"t~aH2?Мp8:!_!,va[&sqSIs:3/ά_?w"XU9eeÆ }pd8W87şwk,!f>ƛ?ɦt,;_&{SXҼ'/ŜwK_*OeAKA>GSY y . =~9a@O/=~qo^J%6\=)SK_9!iK?9! ߇sJWZ rDr^\{c߼bntQ)/WNaxŧZ N5}'栚+BIn|%BwOcah[*,.MzNjŒ{z Wj!_,wϏXHLӿ֢V_^G<_nTR3S܄ `TALwK.L೫|F׸>}c1%x{q,ޮl]%c} bfvaD ¥U|~ #1h.-o_H's_@3a){B*Ah926&3Ư K5!b iǏIdj'1ԠWVAk&8&Zh͘%8ًÄ tH+ 8ӮS]VNvNkפ3 HP|"s݀(]˅AZpb ];)\^?tmю6&;ChqowXum΅\j{}1+Y߱("8Nr`[~ΟjΙ|Q-^Jf I5znŤY;-_D%%~ ,rxJz0 R\ ќoa?{ŕ4sh{uQ(1.\)bp;\[c1E>K0`>27I0jxΔ0f{F9 y*ab/`죗pla"0%M{eR O`CL2cc 8`x0Ia0p~`3 Q pUO%*,]sq;T\n 1X{(!]kb`B Я&A dnhߞC/;F_~:ɬ\w#TEgRH htEG2%yiIg- I"a`r=3TZʹGYRRGaWR>#I9OJb#*3Q{l}r1}L":%@4^?ھGG`P%h.Jtjߨ=3@DъYn=@@vo:B ;Bx݁o^<3o^@=HY03p(,A #7|'7U"ifӸ3p#kٺ/(W%RsCr=\\u '\fX³{K]}1W>x^w ݅^Stdu"cJWMLGBlD%j! ̫y3L3~m)v2>>TNʅ5,yB=/&VtE$w,\>岺LO{I2͞_ UϘO:4ă\@'VZKeѲe|ZybK,3Y|"|u\敇fR >R Hcd)|$sKrll^& +^! /'8҈/9otred14UrbR%D80!JH `҄TE<|v; B8M> @Q"5}wGܶm?Z/oXSTRrW3|}(ggG`\ 7R(a |jac?sוor JAgqQ4)PH0YR3FV~oD֣b1!~o 1.2kTsza1FUtţ4F@ bdinPA۾e!pNjR|6jv׾h/l- uܒT}Q .0IP(.Q5f.ZR)ѯ,gU ?0pՁmwl'?v–}_`qEGށ9Ilm:D@Hx8P*<,Z2d 0\#*];5n"g#{j(*sކnM vxょ6J}-<4~σcݹ] ³q0V}F}7ˁscmFv"tzv,Q@\\az#OkkL__ua t-M"mXi5`ZᎊLN=`^F5#Vo.w)0@}~{~Kk}{vv7ݾ=y?{NRBe;Kܯ``iFIgJL*N*íRi@kV,UA}bن lB.]KWJʅt~Ҡpw*3Lߩ2"0sV/xT'&kVfoYUr8*-V츂(i6Ga!^y{9]syMmY ׫ #ӌ;5zOE_{Mݯd?#ϊUFy6[7L?]lR8oIho\i_kgoq?xJF0f.G8w9fɴP||E![ۥHoŠSG+7yk~-PyPlpz"wxVOPS䤛D|ߨ.sm\D:] B/|v*fRi7}^yvz~ ) Qv0(Mr=ُjyQw;#a5f4 [\Ѱh )T;ڼxzv6?H_3uʗ{gQWݽ*n J[ō{kKy}WUpȮ:A=Dt{X4(=aQ_fwaGuf*j ?ݽU~Pdi wzL] AH;Mtx*(8q'C&GKp/歭I!*OAH j~ι:I۲Sor_QRzzQCRگ J:z0ӪZ1RW`;kJ9r?`nVwuT;_{?b\(+ D svY(-|-խl\;-[nmӗ6^/ͻ_Ervj{߲a3 {[l͜OENش78}{7L~=SYЧc f3p>ƙSsr{s^O_?mw<v}0ܽʿx̲稝M}!ĸuy2U@qXܰ.{|Sh>q?{= '[@lGw m v?ʎgyG'3J; aJ6}mmè7fv+|3@:BRmU!!m|UlOٻ丿"FwW &l tOZGR3S=CCrHhd$^s^]U]F'v^\:k9sKBAzz1K+lLZ\Ib:Bl ÿ}Z9|\ԘsQTJ RߴKKMNܾam?zӎL֧n}E+*9l3P]Yu|/}*lnJ~K-D-&!x& 3D rՐ/9Br&`$?sd'd8,̆S-''>Kb'*?G2 c8KK^WC#>\J^r*}zJ漭RW`JߎZ@mod둯#c#f~v;>պӱr))ƿ>شx|N6Aq v|B){Zͬ6 =;܆S0駻font0>ђU&ȓ&ȉmŐd=,6$A7XtiCMkG18N}74cDo_5Q.Uڅ$ihH219P#E )}?1SX&fI:aF z>jJ 㦼sB̔/_ZB@戺25"e K/uRҙUjM_o6{\̝;[J=,Ϲ͌ ,+`4Paav>-n%kB?"@I,}3GůkU+_mb(ͿkpT : 9g|x@!/Xo>GM$%MM/󵡢\dBY\FP'b'I&('A ~kRm dhٯ#޻<Ί:4 Z3ꗃQDńpC8{? _ (`>Gwvţ95-zYX#͉Y.z`L^J{} [DGVn.Ձ@/ܗZLͤũ}I*0zKҊJF{/uiEDХ]iK\4+mM!fd. 7ٖ@K'`!o3߻A_ qZW@g/)) ҕw:^>fy-)~1Y;㻫]?7l]߱/Mby>UƟ- @8}Ks⻛ݙ57~YceIBELzWmUM"f"F{] %e[GQ]gʟ%\-(󶡙zzj/W)O\{S!Ocֵ$~v?܋Id?krb9UY! GI 3&l"31fj/"P$&TsI;g <0'QdE&Bxg3BЗ(N(Z+!Emp]K3oZei5K>] rM}:hUw9Y|54,k6]]GPV!uX=X_ƶ@2=!m಺K7^QڄX'4e$}~Gw\* 6PSIMfoY ?q<؁by|9Jdžr3bA90dzXvٙZH؈EuЫS_foijze}MHEo&#RđђqsQ@I |º\͔)2aܢ)`'=㶃v;S"Ej$' i9 Lz_Th_(ՊB\/c֡dw%d%Qj0qi_q|/g1P;ŧ|Y!Y66ri vD`Bm:Qj-Kn;,K< ܃Drg+İܙlk+LIy|RbKƸצ˩_̖s_0Wc4xlݪi;w,tXdO?-w*a l(q%EqywnM2_XvAG Iamj젿Iv #zWTĵX%3sMFvƟ ^=$-n)r27۹nxcXf!qVط6-h-gX`BJP{Wڀ.,7-:-5yM³\m 9; u/fkzuNP$ 7=؝m@l{3jRksޏ{M?a11]Bk+ .gH N%t@b5ASt$Gυ;% HPto7g@-Xo5 {^N:m(޺ӖȂ9uAD^ I+ゴ$i0Z<(RH(Gk[gT[2iC[)k#ktHhMc^zFat2aaT.;9!Lu`<\mAjBy{*!/*1$$ R͡oh#hDGWtlz(X9~(ZfBE=%)mY/VOH_Qz[mn$ԡgL(Dr]  Dm3+h(i(c~/o~_zu-s$i`b"@'EKl7%)tR?#.k$JBS6neER|"bƛA}.dL|Ӆ1fQN)ym,ܐw,Pc=e(>ؘ\G#36$~ϒFBΐ3.UhidŇ&mDG쟚֒shNМ{dٶclV3l7 s գ^6\P\QMTgg֝)1!%1T"1Ԅ5@Eo;J8/ m9PӻZl]\-WeD ri{גK9]+o= `05 wƸt¬ͤ8wWO7ku{-M6 qԼa[Z 5vyF_?:f9mMD4Hfnbe /N]/nf+~55./;_TXP"=ZIƱ< KףyO"*TQ]HΥ4ΑRwI4b˹3$!\ {.6+V143%->ۼXvmu+Hlzt.6S$%A02!fClvvD]QO(G i蝳Νu]z-G9Նv}Laԩr;]ۑ@wBfKY+s7=y<+mH oa@ 68 S"bW=!5{8C QOׯjF9mGw`Y&|Y 6UV[Zm5d*4]+(ȓUDjzk)b)=R-V*WZNϦ>+.J(>Ƌ_/;_?Zj(h,4Y@SE)?APzw n/Nn x]~RzҔa^a2S֑%X"1~ v*;sz>%ԽAzTH9 HL`hi~iCYG/v( `RPy\ p[qP7R ".wz )=1qCWa>,fw+\W_O.hx@`%*f  jFjK>`7r6FVy&(f}Ccu` 0-u̟gx5v[@a'ї/oA# uOPe1FR0Ja bs\׊;_C@I 0HY XI$tn87PaI%ŧDEH*HLSX'L `Hb($SJRɮ]۳lC9z 9e FZQG4XF `|~S`:Rq] p0 Ob+-824A:Xb ٳD `F%g=]HdP,} GDJeb$k{B(GBc#Rx'PԈ[IPEQCPXrA ( ~;"돌b؁غ:9ڧx<!Zc-DZU=aH{4m=^*x%Hk=7)R[MgBiQx2e&.p'Rr&xI<)c;;YV8l, >dyi39YP*QMh@\dTa.Fp#.5qQDj{?c .ILt AB? ,JjX8ǀz8B:bhRAVj(͟,k$ZHv~[j_]-GfG~P(b4ɂq m/VJ0&\YΧ(8?yz;e33xjܥH wӅ}Z#˒Iό]sZńb":S0,D>? FjKRX={5GfE($ xVĀ)E.i&ҧk\RɉAՀTGOEFX? OT%8)}6haHpc]Yk@>CQc5qLӻ"W9!]7pKLeD"^_݃^ի&uv>_/d_oN¦Z7YTB1nkg67yxћo֊j"Yht5(T bpuWRmY^}[ 뫲бhaJ%)^mœOLB K*cdYoIa'ÒnmփLY^OV9C9cJOHuJQ:׬e4mYvjuJUwo_ԬO'.|X1'=mˎ~{KDG, TD "TAmJ7],RIh緋O|~,FmyǏ.}@v7E5m3[H|.Vl:?]O[~<ּ7*RNkK?lQv`އݤ3 \gќzm><ۆV<Ⴐʉ U& ַWST䷩#Vǩ@,FE,\"$(25 mUg[yV!UvBβ:Y}k4 C%jS0#.z-V8i׆GU ( o ^QXxITs&ZHDզH.m~ q%~^M(]/O}f1alvifVFZfr%qRacV E2ayH`L;ᝒkb6200*Z9"p`ۚ𭺷@v.F`?j$z@aHR<ư?EbǂE"eF0=,zouк W2éfK.&Ñ}\uVΪ_,C]*q77C>Ւ]F^ׯn;c%Lr_~yt]Mπ%@CO/@Lgc.[uS+!8uo)O7 72͉|O V&g _0V"~!U3R9w|? Aoe@ t>V|>&/L~tRnuBA (D(}ވrC斂Wyg:1AOYʞڎw;G{6k= >c~onQ)G3j@I:9umNՉfvoJ @to00r2 쫟 cKC!u}#Y$,a|oL[%+d3-׼zW{oG[(>{o9]?hO[]V?}p.O!ޢ-Gs߆1 ~EJDAVɇ偹#;Wؗ:xݫi!ğgW+ Aܠ} i QdᤏތW;+A S?@fR#|N7İR4Q֬oz8>-$+Q:UyXm2_dC]QUBW/6ևwv5Nm$fٽQ4%8b֋Uo#K{INϦ>iI〃7s/Fye|ŵ5LjR;3OHq_K4}Hqr'n[ ;R uݰa{p{by$JʁI:>,V2]iZ\xw]|?OWTڀ.}nړ7۠/z\\Yd)/|b0D_(`yx;YiF֟(up0iY[u*Ʊc ȲҷeQ%y7d۴n*Bk*d^YW(˲sVKDeQwLnȫf|g ^L^ℒ2t[̖SE|;kfoMn9-?(p4I[AAa#bT٨#|RxCb!2g+T"ĉzԄSv: Ӊ_qsaqg:PR;wOq尳c9cT=q!Zkȫ±ff][*Fxj|w>5|zu8Dg={j|9NN7(}bZy༬C'Ǜ1RZٚAMLMaHHƠ!,2٧H_ل@*:UUm,[f ds\U[U]U#YS`0M}H1,}o`cԭ;bB3%`Zl%A9#AM Fr0<װ}-V=3N#MnF!4<X%( u]B&y:S ; pŨtz4PQVuSݪtnR鑬Q-Q E@50iy=6Ft6GN@'GF%Gիm z(I[5w^Y\ 2|\r9a,جmͲVЅc@}&wzM? gYzM?hDphD d!cĪ'[u"zߦ:$?hyL{/VNؓmy5'\_5kj7r}h Uߓ-/Rjy^'[xvvvuŵxZI4j65*{-YzE&W%8^úiМD.ذļkD1v$kpZRf7UxY̥6VMI6ԓmy5ꑝ~Z] |: G-9MjQz^5,ף>1|&ZeSVx7PlN7JA v-wݦoDl3ηލxޭөFw*`]ޭT&ӻMa!߸TTabμK|9o8i3 ]C!$}\~ d㽻}s.T|Wm\~.WPŐ c_KKMzA=*z8L7lv=*"ez!o-iUr)ESP1:lr61>0Wqwci7ןw)}< p{JgGfͻ/qoq5JP1KC!.~Op}|TrZLJtxmٿ8x1n:?? > y WO9vW0#}c&%JMsàp ?RonB 8m0g!S(=)whLyi(LLs&&"ubߵQ[PtԢi`Kt(U*}oaγRMna=Oov{VzVGnc7@ZJ/It,5+Ew+a+M|\èXeNi}vV&J ^I:Mn6pқRKn lv;l~-oO~CwgrxeI%1AIj+ - r۰RwYP)5x{qmjV)Ψ ]uspV.+ KNHϿJ]\Vz$Ѷ16]$ :4aÙrLZɶtsm Х`oE%JvL|lZBtېc IaHj'п5Vj{?}1ʌM4&DVĦ^3 B9de,jbդ DIQ. &+fO6EeQbK"&~8!Q%hiqG )*cy6)- /M/3#jv8kNH-FMBM|dVsTĬpquxʸr#Ƒ!Nx)P=Q xyEO)&2\^ѭi:N.fr^E'K ad#)0ivXVJ!V+r:j|@![pMBHjzR?K,f )~㚖ǵs~ԻL),76eQ/,񮖝t!a>4T*y9%*7Rq3 qqڭl1/hnf;7w6 SE4tyvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003627460015134636535017720 0ustar rootrootJan 23 08:51:52 crc systemd[1]: Starting Kubernetes Kubelet... Jan 23 08:51:52 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:52 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 23 08:51:53 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 23 08:51:53 crc kubenswrapper[4899]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.368182 4899 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374880 4899 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374911 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374921 4899 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374930 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374938 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374947 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374955 4899 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374963 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374971 4899 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374979 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374987 4899 feature_gate.go:330] unrecognized feature gate: Example Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.374996 4899 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375005 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375013 4899 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375021 4899 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375029 4899 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375037 4899 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375045 4899 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375053 4899 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375064 4899 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375076 4899 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375085 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375095 4899 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375103 4899 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375112 4899 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375120 4899 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375129 4899 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375138 4899 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375146 4899 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375157 4899 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375167 4899 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375177 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375186 4899 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375195 4899 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375203 4899 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375240 4899 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375248 4899 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375257 4899 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375264 4899 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375272 4899 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375280 4899 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375288 4899 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375297 4899 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375304 4899 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375312 4899 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375320 4899 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375330 4899 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375338 4899 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375346 4899 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375355 4899 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375363 4899 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375371 4899 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375379 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375388 4899 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375396 4899 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375404 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375412 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375422 4899 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375432 4899 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375441 4899 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375449 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375457 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375465 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375473 4899 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375481 4899 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375489 4899 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375498 4899 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375506 4899 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375515 4899 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375525 4899 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.375532 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375673 4899 flags.go:64] FLAG: --address="0.0.0.0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375689 4899 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375705 4899 flags.go:64] FLAG: --anonymous-auth="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375717 4899 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375729 4899 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375738 4899 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375755 4899 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375765 4899 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375776 4899 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375785 4899 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375795 4899 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375806 4899 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375815 4899 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375825 4899 flags.go:64] FLAG: --cgroup-root="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375834 4899 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375843 4899 flags.go:64] FLAG: --client-ca-file="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375852 4899 flags.go:64] FLAG: --cloud-config="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375861 4899 flags.go:64] FLAG: --cloud-provider="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375870 4899 flags.go:64] FLAG: --cluster-dns="[]" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375881 4899 flags.go:64] FLAG: --cluster-domain="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375890 4899 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375899 4899 flags.go:64] FLAG: --config-dir="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375908 4899 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375917 4899 flags.go:64] FLAG: --container-log-max-files="5" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375928 4899 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375937 4899 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375946 4899 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375956 4899 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375966 4899 flags.go:64] FLAG: --contention-profiling="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375974 4899 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375984 4899 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.375993 4899 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376002 4899 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376013 4899 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376023 4899 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376032 4899 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376040 4899 flags.go:64] FLAG: --enable-load-reader="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376049 4899 flags.go:64] FLAG: --enable-server="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376058 4899 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376069 4899 flags.go:64] FLAG: --event-burst="100" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376080 4899 flags.go:64] FLAG: --event-qps="50" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376089 4899 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376098 4899 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376107 4899 flags.go:64] FLAG: --eviction-hard="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376118 4899 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376127 4899 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376137 4899 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376147 4899 flags.go:64] FLAG: --eviction-soft="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376156 4899 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376165 4899 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376174 4899 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376183 4899 flags.go:64] FLAG: --experimental-mounter-path="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376192 4899 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376201 4899 flags.go:64] FLAG: --fail-swap-on="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376237 4899 flags.go:64] FLAG: --feature-gates="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376254 4899 flags.go:64] FLAG: --file-check-frequency="20s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376266 4899 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376278 4899 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376288 4899 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376298 4899 flags.go:64] FLAG: --healthz-port="10248" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376307 4899 flags.go:64] FLAG: --help="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376316 4899 flags.go:64] FLAG: --hostname-override="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376325 4899 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376334 4899 flags.go:64] FLAG: --http-check-frequency="20s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376343 4899 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376352 4899 flags.go:64] FLAG: --image-credential-provider-config="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376361 4899 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376370 4899 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376379 4899 flags.go:64] FLAG: --image-service-endpoint="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376387 4899 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376396 4899 flags.go:64] FLAG: --kube-api-burst="100" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376405 4899 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376415 4899 flags.go:64] FLAG: --kube-api-qps="50" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376424 4899 flags.go:64] FLAG: --kube-reserved="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376433 4899 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376441 4899 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376451 4899 flags.go:64] FLAG: --kubelet-cgroups="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376459 4899 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376469 4899 flags.go:64] FLAG: --lock-file="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376477 4899 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376486 4899 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376495 4899 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376519 4899 flags.go:64] FLAG: --log-json-split-stream="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376533 4899 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376542 4899 flags.go:64] FLAG: --log-text-split-stream="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376551 4899 flags.go:64] FLAG: --logging-format="text" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376560 4899 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376570 4899 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376578 4899 flags.go:64] FLAG: --manifest-url="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376587 4899 flags.go:64] FLAG: --manifest-url-header="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376599 4899 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376608 4899 flags.go:64] FLAG: --max-open-files="1000000" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376619 4899 flags.go:64] FLAG: --max-pods="110" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376628 4899 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376637 4899 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376645 4899 flags.go:64] FLAG: --memory-manager-policy="None" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376655 4899 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376664 4899 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376673 4899 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376682 4899 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376701 4899 flags.go:64] FLAG: --node-status-max-images="50" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376710 4899 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376720 4899 flags.go:64] FLAG: --oom-score-adj="-999" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376729 4899 flags.go:64] FLAG: --pod-cidr="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376738 4899 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376750 4899 flags.go:64] FLAG: --pod-manifest-path="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376759 4899 flags.go:64] FLAG: --pod-max-pids="-1" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376768 4899 flags.go:64] FLAG: --pods-per-core="0" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376777 4899 flags.go:64] FLAG: --port="10250" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376786 4899 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376794 4899 flags.go:64] FLAG: --provider-id="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376805 4899 flags.go:64] FLAG: --qos-reserved="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376814 4899 flags.go:64] FLAG: --read-only-port="10255" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376823 4899 flags.go:64] FLAG: --register-node="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376832 4899 flags.go:64] FLAG: --register-schedulable="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376841 4899 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376855 4899 flags.go:64] FLAG: --registry-burst="10" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376864 4899 flags.go:64] FLAG: --registry-qps="5" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376873 4899 flags.go:64] FLAG: --reserved-cpus="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376883 4899 flags.go:64] FLAG: --reserved-memory="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376893 4899 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376903 4899 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376952 4899 flags.go:64] FLAG: --rotate-certificates="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376962 4899 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376971 4899 flags.go:64] FLAG: --runonce="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376980 4899 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376989 4899 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.376999 4899 flags.go:64] FLAG: --seccomp-default="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377007 4899 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377016 4899 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377026 4899 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377034 4899 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377044 4899 flags.go:64] FLAG: --storage-driver-password="root" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377053 4899 flags.go:64] FLAG: --storage-driver-secure="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377061 4899 flags.go:64] FLAG: --storage-driver-table="stats" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377070 4899 flags.go:64] FLAG: --storage-driver-user="root" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377079 4899 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377088 4899 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377097 4899 flags.go:64] FLAG: --system-cgroups="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377106 4899 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377121 4899 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377129 4899 flags.go:64] FLAG: --tls-cert-file="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377138 4899 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377150 4899 flags.go:64] FLAG: --tls-min-version="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377159 4899 flags.go:64] FLAG: --tls-private-key-file="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377168 4899 flags.go:64] FLAG: --topology-manager-policy="none" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377177 4899 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377186 4899 flags.go:64] FLAG: --topology-manager-scope="container" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377195 4899 flags.go:64] FLAG: --v="2" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377231 4899 flags.go:64] FLAG: --version="false" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377242 4899 flags.go:64] FLAG: --vmodule="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377253 4899 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.377262 4899 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377468 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377478 4899 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377490 4899 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377498 4899 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377507 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377516 4899 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377526 4899 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377537 4899 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377547 4899 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377556 4899 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377564 4899 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377574 4899 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377581 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377589 4899 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377597 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377605 4899 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377613 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377621 4899 feature_gate.go:330] unrecognized feature gate: Example Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377628 4899 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377636 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377644 4899 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377651 4899 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377659 4899 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377667 4899 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377675 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377683 4899 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377691 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377698 4899 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377706 4899 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377714 4899 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377722 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377730 4899 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377738 4899 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377746 4899 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377753 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377766 4899 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377777 4899 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377788 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377797 4899 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377806 4899 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377814 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377823 4899 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377831 4899 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377839 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377847 4899 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377855 4899 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377862 4899 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377871 4899 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377879 4899 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377886 4899 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377894 4899 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377901 4899 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377910 4899 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377917 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377926 4899 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377933 4899 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377941 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377948 4899 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377957 4899 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377964 4899 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377972 4899 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377982 4899 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.377992 4899 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378001 4899 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378011 4899 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378019 4899 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378027 4899 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378038 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378046 4899 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378054 4899 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.378061 4899 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.378086 4899 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.385839 4899 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.385862 4899 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385940 4899 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385947 4899 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385951 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385955 4899 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385959 4899 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385964 4899 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385970 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385974 4899 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385978 4899 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385982 4899 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385985 4899 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385989 4899 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385993 4899 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.385997 4899 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386002 4899 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386006 4899 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386010 4899 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386014 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386018 4899 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386022 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386026 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386029 4899 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386033 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386036 4899 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386040 4899 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386044 4899 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386047 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386051 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386055 4899 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386059 4899 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386062 4899 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386066 4899 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386069 4899 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386073 4899 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386076 4899 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386081 4899 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386085 4899 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386089 4899 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386093 4899 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386096 4899 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386100 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386104 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386107 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386111 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386114 4899 feature_gate.go:330] unrecognized feature gate: Example Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386118 4899 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386122 4899 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386125 4899 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386129 4899 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386132 4899 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386135 4899 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386139 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386143 4899 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386148 4899 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386152 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386156 4899 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386159 4899 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386163 4899 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386166 4899 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386170 4899 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386173 4899 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386177 4899 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386181 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386185 4899 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386188 4899 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386192 4899 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386195 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386199 4899 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386202 4899 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386224 4899 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386227 4899 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.386233 4899 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386338 4899 feature_gate.go:330] unrecognized feature gate: Example Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386343 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386347 4899 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386352 4899 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386356 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386360 4899 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386364 4899 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386369 4899 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386373 4899 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386377 4899 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386381 4899 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386385 4899 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386389 4899 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386393 4899 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386397 4899 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386400 4899 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386404 4899 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386407 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386411 4899 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386415 4899 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386418 4899 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386421 4899 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386425 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386428 4899 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386432 4899 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386435 4899 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386439 4899 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386442 4899 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386446 4899 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386449 4899 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386453 4899 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386457 4899 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386460 4899 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386463 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386468 4899 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386472 4899 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386475 4899 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386479 4899 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386482 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386486 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386489 4899 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386493 4899 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386497 4899 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386501 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386505 4899 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386510 4899 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386514 4899 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386519 4899 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386523 4899 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386527 4899 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386531 4899 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386536 4899 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386541 4899 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386545 4899 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386549 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386554 4899 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386558 4899 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386562 4899 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386566 4899 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386569 4899 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386573 4899 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386576 4899 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386580 4899 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386583 4899 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386587 4899 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386591 4899 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386594 4899 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386599 4899 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386604 4899 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386608 4899 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.386611 4899 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.386617 4899 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.386743 4899 server.go:940] "Client rotation is on, will bootstrap in background" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.388977 4899 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.389041 4899 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.390391 4899 server.go:997] "Starting client certificate rotation" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.390405 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.390634 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 00:31:30.856126739 +0000 UTC Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.390789 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.398464 4899 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.399860 4899 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.400846 4899 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.413188 4899 log.go:25] "Validated CRI v1 runtime API" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.434112 4899 log.go:25] "Validated CRI v1 image API" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.435861 4899 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.437843 4899 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-23-08-46-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.437877 4899 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.454343 4899 manager.go:217] Machine: {Timestamp:2026-01-23 08:51:53.452620567 +0000 UTC m=+0.319590964 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7c35b762-77f6-4705-882d-2520cd67cab9 BootID:173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:48:b3:c2 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:48:b3:c2 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:7e:9e:17 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:7f:4e:a9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:88:8e:7c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8d:fb:d5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5e:99:96:13:b2:25 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5a:88:39:eb:e1:0c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.454828 4899 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.454995 4899 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.455637 4899 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.455825 4899 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.455855 4899 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456082 4899 topology_manager.go:138] "Creating topology manager with none policy" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456096 4899 container_manager_linux.go:303] "Creating device plugin manager" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456387 4899 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456432 4899 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456737 4899 state_mem.go:36] "Initialized new in-memory state store" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.456833 4899 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.457665 4899 kubelet.go:418] "Attempting to sync node with API server" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.457689 4899 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.457716 4899 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.457731 4899 kubelet.go:324] "Adding apiserver pod source" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.457744 4899 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.459735 4899 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.460154 4899 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.461743 4899 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.462713 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.462926 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.462936 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.462950 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.462964 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463000 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463010 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463024 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463033 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463042 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463069 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463079 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463354 4899 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.463949 4899 server.go:1280] "Started kubelet" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.465035 4899 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.465137 4899 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.465405 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.465490 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.465963 4899 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.465992 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:53 crc systemd[1]: Started Kubernetes Kubelet. Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.466071 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.466853 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.466897 4899 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.467084 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 19:31:17.478624888 +0000 UTC Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.469687 4899 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.469943 4899 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.470034 4899 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.470285 4899 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.470336 4899 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.471176 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.471288 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.472377 4899 factory.go:55] Registering systemd factory Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.472437 4899 factory.go:221] Registration of the systemd container factory successfully Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.472452 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="200ms" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.473265 4899 factory.go:153] Registering CRI-O factory Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.473306 4899 factory.go:221] Registration of the crio container factory successfully Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.473400 4899 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.473427 4899 factory.go:103] Registering Raw factory Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.473444 4899 manager.go:1196] Started watching for new ooms in manager Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.475075 4899 manager.go:319] Starting recovery of all containers Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.476587 4899 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188d501cd97239e5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-23 08:51:53.463921125 +0000 UTC m=+0.330891522,LastTimestamp:2026-01-23 08:51:53.463921125 +0000 UTC m=+0.330891522,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.478679 4899 server.go:460] "Adding debug handlers to kubelet server" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486790 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486837 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486848 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486858 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486867 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486877 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486886 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486894 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486905 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486913 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486922 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486932 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486943 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486965 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486977 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486990 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.486998 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487007 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487016 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487025 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487063 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487074 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487086 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487097 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487109 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487143 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487161 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487176 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487192 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487226 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487239 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487261 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487277 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487295 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487307 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487318 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487329 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487341 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487353 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487365 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487377 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487391 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487401 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487413 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487424 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487437 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487447 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487459 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487470 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487480 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487491 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487502 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487523 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487536 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487547 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487559 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487572 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487582 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487592 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487602 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487614 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487625 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487636 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487649 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487661 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487671 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487683 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487692 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487712 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487728 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487744 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487755 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487777 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487790 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487809 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487820 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487831 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487849 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487864 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487876 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487890 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487907 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487918 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487930 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487947 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487958 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487969 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487980 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.487999 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488011 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488022 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488033 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488043 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488054 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488069 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488080 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488089 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488102 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488110 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488118 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488132 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488141 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488149 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488158 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488175 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488186 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488197 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488224 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488234 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488244 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488254 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488269 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488280 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488291 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488300 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488310 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488323 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488334 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488345 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488364 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488380 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488391 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488400 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488408 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488418 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488428 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488439 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488448 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488456 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488464 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488472 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488481 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488488 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488498 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488505 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488514 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488523 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488532 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488541 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488551 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488558 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488567 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488576 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488586 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488594 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488604 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488612 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488621 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488667 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488676 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488686 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.488696 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489362 4899 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489416 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489428 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489437 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489445 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489454 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489462 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489473 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489481 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489490 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489498 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489506 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489515 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489523 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489531 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489539 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489548 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489556 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489566 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489575 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489582 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489591 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489599 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489609 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489618 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489627 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489635 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489644 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489659 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489674 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489683 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489696 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489709 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489719 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489732 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489741 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489750 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489758 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489766 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489783 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489797 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489806 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489814 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489831 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489841 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489851 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489859 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489868 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489876 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489885 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489901 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489910 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489918 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489928 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489937 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489946 4899 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489954 4899 reconstruct.go:97] "Volume reconstruction finished" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.489960 4899 reconciler.go:26] "Reconciler: start to sync state" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.502431 4899 manager.go:324] Recovery completed Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.510722 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.511794 4899 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513262 4899 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513304 4899 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513332 4899 kubelet.go:2335] "Starting kubelet main sync loop" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.513382 4899 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513384 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513448 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.513458 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: W0123 08:51:53.514132 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.514178 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.514269 4899 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.514278 4899 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.514295 4899 state_mem.go:36] "Initialized new in-memory state store" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.570912 4899 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.580383 4899 policy_none.go:49] "None policy: Start" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.581796 4899 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.581839 4899 state_mem.go:35] "Initializing new in-memory state store" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.613550 4899 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.639667 4899 manager.go:334] "Starting Device Plugin manager" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.639982 4899 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.640092 4899 server.go:79] "Starting device plugin registration server" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.640665 4899 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.640757 4899 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.641099 4899 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.641306 4899 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.641384 4899 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.648151 4899 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.673341 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="400ms" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.741929 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.744334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.744394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.744410 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.744447 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.745096 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.813676 4899 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.813782 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.815453 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.815517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.815546 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.815780 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.816495 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.816566 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.816990 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817049 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817240 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817482 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817554 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817940 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817960 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.817972 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818284 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818312 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818405 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818772 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.818834 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819434 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819606 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.819673 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820561 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820619 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820627 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820653 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820920 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.820992 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.822447 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.823162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.823190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.823199 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.894940 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.894996 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895023 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895043 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895062 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895079 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895096 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895114 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895160 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895232 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895282 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895303 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895324 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895346 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.895367 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.945639 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.946848 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.946901 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.946913 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.946940 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:51:53 crc kubenswrapper[4899]: E0123 08:51:53.947445 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996429 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996432 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996694 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996761 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996785 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996865 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996918 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996959 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997017 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996966 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997068 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997102 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997120 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997249 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997187 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997332 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997289 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997390 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.996887 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997434 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997478 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997530 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997575 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997585 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997299 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997673 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997734 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997790 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997528 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 23 08:51:53 crc kubenswrapper[4899]: I0123 08:51:53.997235 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.074680 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="800ms" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.155648 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.163846 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.178502 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.191948 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9bf079e21d531ef1bbe61295b2080123b13f5356178b716c63e5532b2e4b6053 WatchSource:0}: Error finding container 9bf079e21d531ef1bbe61295b2080123b13f5356178b716c63e5532b2e4b6053: Status 404 returned error can't find the container with id 9bf079e21d531ef1bbe61295b2080123b13f5356178b716c63e5532b2e4b6053 Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.192670 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-bfba6bb1824798fc748a5cc7afb80649f640cc21ce00eedb1c3fc66ab3c7e51e WatchSource:0}: Error finding container bfba6bb1824798fc748a5cc7afb80649f640cc21ce00eedb1c3fc66ab3c7e51e: Status 404 returned error can't find the container with id bfba6bb1824798fc748a5cc7afb80649f640cc21ce00eedb1c3fc66ab3c7e51e Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.199435 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-bcba16ee0bcabf41309335c855cd7b04f78e64f9bb9f6b7b9d1b999f4ffd576d WatchSource:0}: Error finding container bcba16ee0bcabf41309335c855cd7b04f78e64f9bb9f6b7b9d1b999f4ffd576d: Status 404 returned error can't find the container with id bcba16ee0bcabf41309335c855cd7b04f78e64f9bb9f6b7b9d1b999f4ffd576d Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.199564 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.208854 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.212893 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-910f96258e2ebb46fb25d958938a4a0b34e7d1f3b679d496a97f66584a9a0954 WatchSource:0}: Error finding container 910f96258e2ebb46fb25d958938a4a0b34e7d1f3b679d496a97f66584a9a0954: Status 404 returned error can't find the container with id 910f96258e2ebb46fb25d958938a4a0b34e7d1f3b679d496a97f66584a9a0954 Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.239098 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ff9251aed05889d54aa80fdce99893e3eecba994eaf73920f6c0d047f806155b WatchSource:0}: Error finding container ff9251aed05889d54aa80fdce99893e3eecba994eaf73920f6c0d047f806155b: Status 404 returned error can't find the container with id ff9251aed05889d54aa80fdce99893e3eecba994eaf73920f6c0d047f806155b Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.348481 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.351120 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.351186 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.351199 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.351257 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.352323 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.469230 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 00:14:02.274277629 +0000 UTC Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.472089 4899 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.517456 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bfba6bb1824798fc748a5cc7afb80649f640cc21ce00eedb1c3fc66ab3c7e51e"} Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.518663 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ff9251aed05889d54aa80fdce99893e3eecba994eaf73920f6c0d047f806155b"} Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.519533 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"910f96258e2ebb46fb25d958938a4a0b34e7d1f3b679d496a97f66584a9a0954"} Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.520231 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bcba16ee0bcabf41309335c855cd7b04f78e64f9bb9f6b7b9d1b999f4ffd576d"} Jan 23 08:51:54 crc kubenswrapper[4899]: I0123 08:51:54.521252 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9bf079e21d531ef1bbe61295b2080123b13f5356178b716c63e5532b2e4b6053"} Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.544012 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.544104 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.545590 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.545627 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.557198 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.557300 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:54 crc kubenswrapper[4899]: W0123 08:51:54.866881 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.867014 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:54 crc kubenswrapper[4899]: E0123 08:51:54.875853 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="1.6s" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.152447 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.153607 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.153642 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.153652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.153675 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:51:55 crc kubenswrapper[4899]: E0123 08:51:55.154233 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.452272 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 23 08:51:55 crc kubenswrapper[4899]: E0123 08:51:55.453516 4899 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.470114 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 13:11:56.762338993 +0000 UTC Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.471051 4899 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.525573 4899 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="23dbb59d872d663a876e9d320bd90e3ee539f107377299991998b285a63b10b8" exitCode=0 Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.525636 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"23dbb59d872d663a876e9d320bd90e3ee539f107377299991998b285a63b10b8"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.525649 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.527530 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.527572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.527585 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.530420 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.530449 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.530461 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.530463 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.530470 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.531600 4899 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9" exitCode=0 Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.531656 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.531697 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532385 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532636 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532509 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532923 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834" exitCode=0 Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532947 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.532974 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.533007 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.533014 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.533686 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.533712 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.533723 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.534539 4899 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247" exitCode=0 Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.534588 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.534602 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247"} Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.535134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.535157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.535165 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.538682 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.541438 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.541579 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:55 crc kubenswrapper[4899]: I0123 08:51:55.541609 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.470849 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 04:57:08.292185642 +0000 UTC Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.539251 4899 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f736c2fc78cb296e65f67467c6cce91ce967652ae2b67eebf638d999ff3178eb" exitCode=0 Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.539320 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f736c2fc78cb296e65f67467c6cce91ce967652ae2b67eebf638d999ff3178eb"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.539408 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.540827 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.540859 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.540871 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.548225 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.548268 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.548282 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.548335 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.549514 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.549550 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.549559 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552380 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552419 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552432 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552441 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552451 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.552466 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.556550 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.556695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.556818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.560030 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.560301 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd"} Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.560493 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.561643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.561761 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.561846 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.562945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.563067 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.563162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.755115 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.756645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.756689 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.756698 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:56 crc kubenswrapper[4899]: I0123 08:51:56.756773 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.322229 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.327349 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.471935 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 18:37:29.588664521 +0000 UTC Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.503719 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.568779 4899 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="83bc211613c8889b1c214d3503fb83ee7b0ab2f0543fcd95b784ebe4b601d86b" exitCode=0 Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.568857 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"83bc211613c8889b1c214d3503fb83ee7b0ab2f0543fcd95b784ebe4b601d86b"} Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.568958 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.569000 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.569076 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.569100 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.568966 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.569474 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.569428 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.570688 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.570747 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.570781 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.571464 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.571500 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.571511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.571948 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.571992 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572330 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572347 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572357 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572470 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572500 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.572516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:57 crc kubenswrapper[4899]: I0123 08:51:57.630322 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.473177 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 12:34:25.097360138 +0000 UTC Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.485454 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577326 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd057062f7680519fe8fee23eabb0d9994b79474b9c46e3293cd79220d95684e"} Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577407 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577422 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577460 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577407 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577500 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.577414 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b4c1a320817528fef7347af28c56b752dd79cd884c15441aa33bfedea374b48c"} Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.578315 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"54e65a53617c7eeeb2bbd234503ed2546384859d8fcc727f11e26db2744b100f"} Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.578363 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de86be5abee9338d5508816dc4f77e379f1b14a70a4254e79174c608e282a242"} Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.578382 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b26664c1dbf488573e455beba962306bf6750a2d359e4b3d222b24f7816331cc"} Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.580786 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.580915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.581001 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582054 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582080 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582090 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582242 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582286 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582301 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582659 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:58 crc kubenswrapper[4899]: I0123 08:51:58.582668 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.473333 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 09:42:30.967804309 +0000 UTC Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.580167 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.580562 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.580424 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.580238 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582129 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582142 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582096 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582613 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582521 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.582944 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.583068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:51:59 crc kubenswrapper[4899]: I0123 08:51:59.854659 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 23 08:52:00 crc kubenswrapper[4899]: I0123 08:52:00.474171 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 07:28:46.964668944 +0000 UTC Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.049352 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.049512 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.050932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.050957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.050965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.141703 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.474909 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 17:59:54.725893331 +0000 UTC Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.486276 4899 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.486406 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.586083 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.586999 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.587040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:01 crc kubenswrapper[4899]: I0123 08:52:01.587049 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.118039 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.118326 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.120050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.120104 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.120122 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:02 crc kubenswrapper[4899]: I0123 08:52:02.475407 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 21:29:17.9725997 +0000 UTC Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.476591 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 14:49:25.488619734 +0000 UTC Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.636116 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.636304 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.637781 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.637826 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:03 crc kubenswrapper[4899]: I0123 08:52:03.637838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:03 crc kubenswrapper[4899]: E0123 08:52:03.648524 4899 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 23 08:52:04 crc kubenswrapper[4899]: I0123 08:52:04.477021 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 14:31:24.738110052 +0000 UTC Jan 23 08:52:05 crc kubenswrapper[4899]: I0123 08:52:05.477539 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 07:46:05.532516101 +0000 UTC Jan 23 08:52:06 crc kubenswrapper[4899]: I0123 08:52:06.472412 4899 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 23 08:52:06 crc kubenswrapper[4899]: I0123 08:52:06.477671 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 22:53:49.087655188 +0000 UTC Jan 23 08:52:06 crc kubenswrapper[4899]: E0123 08:52:06.477696 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Jan 23 08:52:06 crc kubenswrapper[4899]: W0123 08:52:06.517136 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 23 08:52:06 crc kubenswrapper[4899]: I0123 08:52:06.517260 4899 trace.go:236] Trace[915809340]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Jan-2026 08:51:56.515) (total time: 10001ms): Jan 23 08:52:06 crc kubenswrapper[4899]: Trace[915809340]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:52:06.517) Jan 23 08:52:06 crc kubenswrapper[4899]: Trace[915809340]: [10.001498545s] [10.001498545s] END Jan 23 08:52:06 crc kubenswrapper[4899]: E0123 08:52:06.517291 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 23 08:52:06 crc kubenswrapper[4899]: E0123 08:52:06.757479 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Jan 23 08:52:06 crc kubenswrapper[4899]: W0123 08:52:06.790059 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 23 08:52:06 crc kubenswrapper[4899]: I0123 08:52:06.790340 4899 trace.go:236] Trace[426277716]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Jan-2026 08:51:56.788) (total time: 10001ms): Jan 23 08:52:06 crc kubenswrapper[4899]: Trace[426277716]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:52:06.790) Jan 23 08:52:06 crc kubenswrapper[4899]: Trace[426277716]: [10.001367892s] [10.001367892s] END Jan 23 08:52:06 crc kubenswrapper[4899]: E0123 08:52:06.790485 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 23 08:52:07 crc kubenswrapper[4899]: W0123 08:52:07.069118 4899 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.069239 4899 trace.go:236] Trace[1922735991]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Jan-2026 08:51:57.067) (total time: 10001ms): Jan 23 08:52:07 crc kubenswrapper[4899]: Trace[1922735991]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:52:07.069) Jan 23 08:52:07 crc kubenswrapper[4899]: Trace[1922735991]: [10.001613821s] [10.001613821s] END Jan 23 08:52:07 crc kubenswrapper[4899]: E0123 08:52:07.069264 4899 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.192650 4899 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.192978 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.209449 4899 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.209799 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.478153 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 12:59:05.714417296 +0000 UTC Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.511709 4899 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]log ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]etcd ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/generic-apiserver-start-informers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/priority-and-fairness-filter ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-apiextensions-informers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-apiextensions-controllers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/crd-informer-synced ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-system-namespaces-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 23 08:52:07 crc kubenswrapper[4899]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/bootstrap-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/start-kube-aggregator-informers ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-registration-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-discovery-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]autoregister-completion ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-openapi-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 23 08:52:07 crc kubenswrapper[4899]: livez check failed Jan 23 08:52:07 crc kubenswrapper[4899]: I0123 08:52:07.511781 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:52:08 crc kubenswrapper[4899]: I0123 08:52:08.478988 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 17:28:48.200868012 +0000 UTC Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.480145 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 18:15:37.256705775 +0000 UTC Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.957948 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.959108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.959142 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.959154 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:09 crc kubenswrapper[4899]: I0123 08:52:09.959177 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:52:09 crc kubenswrapper[4899]: E0123 08:52:09.963005 4899 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 23 08:52:10 crc kubenswrapper[4899]: I0123 08:52:10.481034 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 04:24:08.155930102 +0000 UTC Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.076253 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.076446 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.077653 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.077694 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.077710 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.094149 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.481816 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 16:28:01.608318661 +0000 UTC Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.487049 4899 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.487108 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.613646 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.615307 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.615382 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:11 crc kubenswrapper[4899]: I0123 08:52:11.615406 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.091311 4899 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.123940 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.124109 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.125379 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.125423 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.125438 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.188545 4899 trace.go:236] Trace[1084766994]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Jan-2026 08:51:57.205) (total time: 14983ms): Jan 23 08:52:12 crc kubenswrapper[4899]: Trace[1084766994]: ---"Objects listed" error: 14983ms (08:52:12.188) Jan 23 08:52:12 crc kubenswrapper[4899]: Trace[1084766994]: [14.983139503s] [14.983139503s] END Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.188572 4899 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.190304 4899 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.198270 4899 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.319574 4899 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.482088 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 13:11:42.723259758 +0000 UTC Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.508921 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.509095 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.509685 4899 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.509760 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.510235 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.510305 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.510325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.513026 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.615620 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.617340 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a" exitCode=255 Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.617395 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a"} Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.617560 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.618631 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.618696 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.618757 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:12 crc kubenswrapper[4899]: I0123 08:52:12.619989 4899 scope.go:117] "RemoveContainer" containerID="e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.122023 4899 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.468072 4899 apiserver.go:52] "Watching apiserver" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.471716 4899 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.472023 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.472660 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.472719 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.472845 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.472843 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.472906 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.473091 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.473133 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.473405 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.473505 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475134 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475159 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475184 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475511 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475514 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.475597 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.478229 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.478288 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.481108 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.482319 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 15:28:36.990131497 +0000 UTC Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.502424 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.513180 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.526120 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.536495 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.545541 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.552773 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.560557 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.568889 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.571401 4899 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.579816 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.589372 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.598476 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599842 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599884 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599910 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599934 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599958 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.599979 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600001 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600023 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600053 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600074 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600096 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600116 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600166 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600186 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600221 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600260 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600551 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600582 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600600 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600686 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600804 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600823 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600868 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600895 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600940 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601055 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600968 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.600973 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601026 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601081 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601033 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601121 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601147 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601172 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601238 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601246 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601261 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601286 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601306 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601328 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601350 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601370 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601390 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601410 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601432 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601453 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601473 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601491 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601507 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601524 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601540 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601554 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601570 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601587 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601628 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601644 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601660 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601674 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601701 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601716 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601260 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601305 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601491 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601746 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601771 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601798 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601821 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601844 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601867 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601890 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601912 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601934 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601957 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601979 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602002 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602024 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602046 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602104 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602129 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602155 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602178 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602203 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602243 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602267 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602294 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602318 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602342 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602372 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602452 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602478 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602507 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602539 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602565 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602592 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602614 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602636 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602661 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602681 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602701 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602759 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602783 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602804 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602821 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602836 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602855 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602876 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602895 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602916 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602936 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602960 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602981 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603002 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603024 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601517 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603045 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601631 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601711 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601738 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601803 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603078 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603107 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603133 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601936 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.601998 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602004 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602184 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602195 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602238 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603225 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603254 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603278 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603299 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603324 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603347 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603368 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603389 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603412 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603433 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603451 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603473 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603493 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603515 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603536 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603561 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603584 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603605 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603626 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603648 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603671 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603691 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603713 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603736 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603758 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603782 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603803 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603824 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603847 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603868 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603888 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603918 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603945 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603978 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604002 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604025 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604047 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604069 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604093 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604116 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604147 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604175 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604197 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604245 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604271 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604292 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604316 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604339 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604518 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604541 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604566 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604588 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604609 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604632 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604656 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604676 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604693 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604710 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604726 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604742 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604759 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604773 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604788 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604804 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604825 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604842 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604858 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604874 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604889 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604905 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604920 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604936 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604952 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604967 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604982 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.604997 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605013 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605030 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605046 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605063 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605079 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605096 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605114 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605129 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605145 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605160 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605177 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605198 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605241 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605262 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605287 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605309 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605325 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605347 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605374 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605430 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605461 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605494 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605522 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605550 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605578 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605606 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605630 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605656 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605677 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605697 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605720 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605758 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605784 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605846 4899 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605860 4899 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605875 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605888 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605901 4899 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605915 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605929 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605942 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605954 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605967 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605980 4899 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605994 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606006 4899 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606019 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606032 4899 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606045 4899 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606058 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606070 4899 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606082 4899 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606094 4899 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606106 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606119 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606132 4899 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606145 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606158 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606171 4899 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606185 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606198 4899 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602276 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602405 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602434 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602604 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602686 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602741 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.602782 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603000 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603021 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.603030 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605459 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.605527 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606284 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606423 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606678 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606763 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606759 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.606975 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607065 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607256 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607381 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607470 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607621 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607631 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607874 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607891 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.607950 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608067 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608247 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608496 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608733 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608764 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.608914 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609020 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609179 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609462 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609507 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609477 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609820 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609839 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609849 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.609966 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.610286 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:14.11026496 +0000 UTC m=+20.977235347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.610729 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.610995 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611252 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611331 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611433 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611568 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611801 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611819 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611830 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611866 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.611921 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612114 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612129 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612407 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612465 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612555 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612773 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.612928 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.613229 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.614276 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.614313 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.614395 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.614786 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.614858 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.615270 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.615834 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.616160 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.616626 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.617057 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.617376 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.617746 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.618012 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.618409 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.618676 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.618808 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.619439 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.619534 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.619281 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.619796 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.620153 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.620465 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.620592 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.620593 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.622235 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.622720 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.623022 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.623256 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.623792 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.623953 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.623992 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.624496 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.624789 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.624920 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.624983 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.624940 4899 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.625643 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.626278 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.626883 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.626986 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.627020 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.627137 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.627855 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.627840 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.627941 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:14.127911616 +0000 UTC m=+20.994882013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.627992 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.628015 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.628157 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.628342 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.628550 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.629106 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.629277 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.629423 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.629705 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.630763 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.631076 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.631652 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.632027 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.633295 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.634338 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.633890 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.634607 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:14.134439068 +0000 UTC m=+21.001409455 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.634867 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635280 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635410 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635588 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635595 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635681 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635727 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635811 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.635822 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.636397 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.637984 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.638587 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.639582 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.639920 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.640460 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.640251 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.641459 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.642647 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.643498 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.644059 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.644397 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.644780 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.643643 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.644983 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.645507 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.646404 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.646922 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.647492 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.647757 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.647888 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.648297 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.648305 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.648339 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.648705 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.648844 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.649841 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650072 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650275 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650413 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650703 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:14.150644377 +0000 UTC m=+21.017614994 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.650831 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.650323 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650131 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650886 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650900 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.650948 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:14.150932094 +0000 UTC m=+21.017902471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.651533 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.651566 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.651281 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.652698 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.654888 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.655529 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.659116 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.659305 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.659347 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.659426 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.660004 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" exitCode=255 Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.660049 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548"} Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.660060 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.660101 4899 scope.go:117] "RemoveContainer" containerID="e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.663127 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.663137 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.663688 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.664122 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.668707 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.668902 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.669281 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.675091 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.675193 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.675398 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:52:13 crc kubenswrapper[4899]: E0123 08:52:13.675631 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.676304 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.681538 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.687779 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.692706 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.697071 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.699882 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707492 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707524 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707599 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707603 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707638 4899 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707644 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707649 4899 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707675 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707687 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707699 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707711 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707722 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707735 4899 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707746 4899 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707757 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707768 4899 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707781 4899 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707791 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707802 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707813 4899 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707823 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707834 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707846 4899 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707858 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707870 4899 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707880 4899 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707892 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707903 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707913 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707925 4899 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707936 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707949 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707961 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707984 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.707996 4899 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708007 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708018 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708029 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708040 4899 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708051 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708061 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708073 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708084 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708094 4899 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708106 4899 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708116 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708128 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708140 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708152 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708163 4899 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708174 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708185 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708196 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708225 4899 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708237 4899 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708250 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708262 4899 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708273 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708284 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708297 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708316 4899 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708329 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708341 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708354 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708365 4899 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708377 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708389 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708400 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708412 4899 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708424 4899 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708436 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708447 4899 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708460 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708472 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708485 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708497 4899 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708514 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708527 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708539 4899 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708550 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708562 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708573 4899 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708599 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708610 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708623 4899 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708663 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708675 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708688 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708700 4899 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708712 4899 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708723 4899 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708735 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708747 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708758 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708769 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708782 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708794 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708806 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708819 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708831 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708844 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708857 4899 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708869 4899 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708880 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708899 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708911 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708922 4899 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708934 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708945 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708964 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708976 4899 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.708989 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709008 4899 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709019 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709033 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709044 4899 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709055 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709067 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709078 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709092 4899 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709104 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709116 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709128 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709140 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709150 4899 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709161 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709172 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709184 4899 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709195 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709206 4899 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709236 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709248 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709260 4899 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709272 4899 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709284 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709295 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709307 4899 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709319 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709332 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709343 4899 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709355 4899 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709367 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709378 4899 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709390 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709401 4899 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709413 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709425 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709437 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709448 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709460 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709471 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709482 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709493 4899 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709505 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709516 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709527 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709538 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709549 4899 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709561 4899 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709572 4899 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709584 4899 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709596 4899 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709608 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709619 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709633 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709645 4899 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709656 4899 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709668 4899 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709680 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.709692 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.712891 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.722311 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.731902 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.787444 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.794600 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 23 08:52:13 crc kubenswrapper[4899]: W0123 08:52:13.798733 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-57823f9a1c2203680d957b54a94ebc8162400373f405ea482f5733dcc35b2f7d WatchSource:0}: Error finding container 57823f9a1c2203680d957b54a94ebc8162400373f405ea482f5733dcc35b2f7d: Status 404 returned error can't find the container with id 57823f9a1c2203680d957b54a94ebc8162400373f405ea482f5733dcc35b2f7d Jan 23 08:52:13 crc kubenswrapper[4899]: I0123 08:52:13.799437 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 23 08:52:13 crc kubenswrapper[4899]: W0123 08:52:13.810049 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-92914a03fb9e0511a65da866b3b851f719ad25278a7585742cabbfe1290be74e WatchSource:0}: Error finding container 92914a03fb9e0511a65da866b3b851f719ad25278a7585742cabbfe1290be74e: Status 404 returned error can't find the container with id 92914a03fb9e0511a65da866b3b851f719ad25278a7585742cabbfe1290be74e Jan 23 08:52:13 crc kubenswrapper[4899]: W0123 08:52:13.812117 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4a108c0afdc7582f8a9cc83b66ff3f797e711f51f4c145f65268ee4ff7a9c954 WatchSource:0}: Error finding container 4a108c0afdc7582f8a9cc83b66ff3f797e711f51f4c145f65268ee4ff7a9c954: Status 404 returned error can't find the container with id 4a108c0afdc7582f8a9cc83b66ff3f797e711f51f4c145f65268ee4ff7a9c954 Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.112307 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.112458 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:15.112435469 +0000 UTC m=+21.979405856 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.213689 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.213731 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.213754 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.213773 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213818 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213871 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213911 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:15.21389119 +0000 UTC m=+22.080861587 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213932 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:15.213922491 +0000 UTC m=+22.080892898 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213933 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213964 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.214013 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.214036 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.213969 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.214106 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.214089 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:15.214072094 +0000 UTC m=+22.081042511 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.214177 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:15.214162417 +0000 UTC m=+22.081132804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.482542 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 04:49:43.452141315 +0000 UTC Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.663673 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.663724 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"57823f9a1c2203680d957b54a94ebc8162400373f405ea482f5733dcc35b2f7d"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.665298 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.667284 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:52:14 crc kubenswrapper[4899]: E0123 08:52:14.667494 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.668334 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.668376 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.668395 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a108c0afdc7582f8a9cc83b66ff3f797e711f51f4c145f65268ee4ff7a9c954"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.669327 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"92914a03fb9e0511a65da866b3b851f719ad25278a7585742cabbfe1290be74e"} Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.676451 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2bdbd7f6f65da8e9f0f95c5f6dd6066bac9e7d5ab0dc404a2f72fe450082f4a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:12Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0123 08:52:06.782306 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:06.783859 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2651750563/tls.crt::/tmp/serving-cert-2651750563/tls.key\\\\\\\"\\\\nI0123 08:52:12.215125 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:12.217629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:12.217694 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:12.217752 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:12.217786 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:12.223605 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:12.223651 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:12.223657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:12.223663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:12.223668 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:12.223672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:12.223677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:12.223607 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:12.226107 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.689333 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.698697 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.708431 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.720520 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.734014 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.749050 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.761776 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.772680 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.786432 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.797352 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.806759 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.816386 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.825329 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:14 crc kubenswrapper[4899]: I0123 08:52:14.987902 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.120985 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.121297 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:17.121254786 +0000 UTC m=+23.988225213 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.221608 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.221672 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.221695 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.221720 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221775 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221808 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221827 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221835 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:17.221818964 +0000 UTC m=+24.088789351 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221839 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221873 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:17.221862635 +0000 UTC m=+24.088833022 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221926 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221939 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221948 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.221974 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:17.221966198 +0000 UTC m=+24.088936585 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.222025 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.222050 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:17.22204279 +0000 UTC m=+24.089013177 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.482841 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 02:39:48.807779399 +0000 UTC Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.514398 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.514409 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.514536 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.514616 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.514424 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.514690 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.519258 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.519988 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.521256 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.521956 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.523061 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.523606 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.524235 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.525089 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.525692 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.526632 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.527248 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.528464 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.528994 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.529632 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.530646 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.531281 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.532413 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.532885 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.533526 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.534721 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.535262 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.536401 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.536861 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.537828 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.538256 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.538948 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.540134 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.540582 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.541645 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.542265 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.543300 4899 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.543478 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.545154 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.545665 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.546094 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.547445 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.548117 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.548706 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.549422 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.550156 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.550638 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.551310 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.551961 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.552613 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.553051 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.556346 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.556813 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.557969 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.558482 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.559305 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.559742 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.560220 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.561094 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.561647 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.672516 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:52:15 crc kubenswrapper[4899]: E0123 08:52:15.672735 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 23 08:52:15 crc kubenswrapper[4899]: I0123 08:52:15.949271 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.363922 4899 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.365517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.365547 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.365555 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.365616 4899 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.378856 4899 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.379086 4899 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.380349 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.380400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.380411 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.380428 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.380440 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.397882 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.401098 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.401145 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.401156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.401175 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.401186 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.415035 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.419123 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.419165 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.419177 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.419193 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.419224 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.434698 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.439455 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.439491 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.439501 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.439517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.439527 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.452809 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.457107 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.457136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.457146 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.457162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.457173 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.469985 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.470108 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.471754 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.471795 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.471806 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.471822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.471832 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.483237 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 08:09:10.873210119 +0000 UTC Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.574510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.574555 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.574566 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.574581 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.574591 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.675768 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.676527 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:52:16 crc kubenswrapper[4899]: E0123 08:52:16.676684 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.677099 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.677144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.677161 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.677190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.677224 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.694763 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.707972 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.724908 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.736300 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.748533 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.768950 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.779558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.779614 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.779625 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.779643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.779654 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.782248 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:16Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.881448 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.881492 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.881514 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.881534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.881548 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.984367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.984408 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.984417 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.984430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:16 crc kubenswrapper[4899]: I0123 08:52:16.984439 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:16Z","lastTransitionTime":"2026-01-23T08:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.087050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.087106 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.087118 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.087137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.087150 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.137573 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.137792 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:21.137754533 +0000 UTC m=+28.004724950 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.189387 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.189441 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.189453 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.189471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.189483 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.238998 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.239037 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.239055 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.239072 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239158 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239202 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:21.239189903 +0000 UTC m=+28.106160290 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239382 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239406 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239420 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239455 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:21.23944285 +0000 UTC m=+28.106413247 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239450 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239509 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239520 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239528 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239545 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:21.239523972 +0000 UTC m=+28.106494429 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.239574 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:21.239564343 +0000 UTC m=+28.106534820 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.292540 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.292577 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.292587 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.292603 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.292615 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.395066 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.395103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.395111 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.395126 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.395137 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.477687 4899 csr.go:261] certificate signing request csr-vjtzg is approved, waiting to be issued Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.483896 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 00:58:39.506731542 +0000 UTC Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.484375 4899 csr.go:257] certificate signing request csr-vjtzg is issued Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.497196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.497259 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.497276 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.497291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.497301 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.513807 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.513830 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.513925 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.514091 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.514190 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:17 crc kubenswrapper[4899]: E0123 08:52:17.514308 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.600198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.600254 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.600265 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.600281 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.600291 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.625306 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-nzqf7"] Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.625588 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.629246 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.629336 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.630165 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.642490 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.653355 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.665514 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.683060 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.692466 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.702589 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.702739 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.702787 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.702815 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.702831 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.707180 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.719422 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.742612 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:17Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.742925 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7629148-8035-4115-b7f7-89c2868cad0c-hosts-file\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.742993 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjqs7\" (UniqueName: \"kubernetes.io/projected/d7629148-8035-4115-b7f7-89c2868cad0c-kube-api-access-sjqs7\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.805196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.805248 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.805256 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.805269 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.805277 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.843764 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjqs7\" (UniqueName: \"kubernetes.io/projected/d7629148-8035-4115-b7f7-89c2868cad0c-kube-api-access-sjqs7\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.843827 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7629148-8035-4115-b7f7-89c2868cad0c-hosts-file\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.843881 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7629148-8035-4115-b7f7-89c2868cad0c-hosts-file\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.864891 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjqs7\" (UniqueName: \"kubernetes.io/projected/d7629148-8035-4115-b7f7-89c2868cad0c-kube-api-access-sjqs7\") pod \"node-resolver-nzqf7\" (UID: \"d7629148-8035-4115-b7f7-89c2868cad0c\") " pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.908240 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.908303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.908316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.908337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.908350 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:17Z","lastTransitionTime":"2026-01-23T08:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.936270 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nzqf7" Jan 23 08:52:17 crc kubenswrapper[4899]: W0123 08:52:17.947296 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7629148_8035_4115_b7f7_89c2868cad0c.slice/crio-63c959ed0e363d44d81089eb8ebc0e4176974e7fd741197e19c06f3fd9952d8c WatchSource:0}: Error finding container 63c959ed0e363d44d81089eb8ebc0e4176974e7fd741197e19c06f3fd9952d8c: Status 404 returned error can't find the container with id 63c959ed0e363d44d81089eb8ebc0e4176974e7fd741197e19c06f3fd9952d8c Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.986980 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h5gln"] Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.987349 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h5gln" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.989540 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.989560 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.989724 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.990713 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 23 08:52:17 crc kubenswrapper[4899]: I0123 08:52:17.994288 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.006468 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.010472 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.010503 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.010511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.010526 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.010536 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.021381 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.041429 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.044856 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.044910 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-os-release\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.044931 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-k8s-cni-cncf-io\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.044954 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp9g8\" (UniqueName: \"kubernetes.io/projected/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-kube-api-access-sp9g8\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.044980 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-netns\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045047 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-kubelet\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045094 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-daemon-config\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045124 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cnibin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045181 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-system-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045239 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-bin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045266 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-multus\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045291 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-conf-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045328 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-hostroot\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045381 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-etc-kubernetes\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045417 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cni-binary-copy\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045442 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-socket-dir-parent\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.045527 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-multus-certs\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.057385 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.071936 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.085733 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.099156 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.112736 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.112784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.112795 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.112813 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.112823 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.117278 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.137988 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147544 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-hostroot\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147606 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-etc-kubernetes\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147637 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cni-binary-copy\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147660 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-socket-dir-parent\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147680 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-multus-certs\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147710 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-hostroot\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147731 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-etc-kubernetes\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147729 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147796 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-multus-certs\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147793 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-socket-dir-parent\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147846 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-os-release\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147876 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-k8s-cni-cncf-io\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147894 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp9g8\" (UniqueName: \"kubernetes.io/projected/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-kube-api-access-sp9g8\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147910 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147918 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-os-release\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147942 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-kubelet\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147967 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-daemon-config\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147993 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cnibin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148013 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-netns\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.147996 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-k8s-cni-cncf-io\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148041 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-kubelet\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148031 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-system-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148116 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-system-cni-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148125 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cnibin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148151 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-bin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148131 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-bin\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148123 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-run-netns\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148195 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-multus\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148241 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-host-var-lib-cni-multus\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148247 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-conf-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148330 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-conf-dir\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148566 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-cni-binary-copy\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.148910 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-multus-daemon-config\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.162924 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp9g8\" (UniqueName: \"kubernetes.io/projected/60a84ebf-a188-4eee-8a9f-e02ac7adb8c8-kube-api-access-sp9g8\") pod \"multus-h5gln\" (UID: \"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\") " pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.215103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.215151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.215163 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.215186 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.215198 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.306506 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h5gln" Jan 23 08:52:18 crc kubenswrapper[4899]: W0123 08:52:18.316280 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60a84ebf_a188_4eee_8a9f_e02ac7adb8c8.slice/crio-55bc1bd4cb6f1b20ea152f063b261f9801c8b69c055dedb9511ff5a1def9a942 WatchSource:0}: Error finding container 55bc1bd4cb6f1b20ea152f063b261f9801c8b69c055dedb9511ff5a1def9a942: Status 404 returned error can't find the container with id 55bc1bd4cb6f1b20ea152f063b261f9801c8b69c055dedb9511ff5a1def9a942 Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.317921 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.317953 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.317965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.317982 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.317992 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.364013 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gbqr2"] Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.364746 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.366456 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qmqs"] Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.367083 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.367354 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.370111 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.370497 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.370530 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.370835 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.372321 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.372387 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.372634 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.374523 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.374820 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-x6wwr"] Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.375314 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.378964 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.379288 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.379499 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.379946 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.381792 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.391900 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.414926 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.421592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.421642 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.421658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.421679 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.421693 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.431148 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.448100 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451673 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb8vr\" (UniqueName: \"kubernetes.io/projected/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-kube-api-access-jb8vr\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451724 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451747 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cnibin\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451789 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451848 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451942 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451966 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-system-cni-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.451988 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452011 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452093 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452143 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452167 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxsd6\" (UniqueName: \"kubernetes.io/projected/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-kube-api-access-zxsd6\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452189 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452242 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452268 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mv9\" (UniqueName: \"kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452289 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-rootfs\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452314 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-os-release\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452454 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452524 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452551 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-proxy-tls\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452572 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452587 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452614 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452636 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-binary-copy\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452670 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452702 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452721 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452770 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452790 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452812 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.452853 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.470579 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.485025 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 14:32:38.541036416 +0000 UTC Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.486106 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-23 08:47:17 +0000 UTC, rotation deadline is 2026-10-27 18:34:03.813447553 +0000 UTC Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.486144 4899 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6657h41m45.32730658s for next certificate rotation Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.487992 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.495672 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.500383 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.506542 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.517171 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.524385 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.524430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.524443 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.524471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.524491 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.531248 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.539449 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.548767 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553647 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb8vr\" (UniqueName: \"kubernetes.io/projected/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-kube-api-access-jb8vr\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553706 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553731 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cnibin\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553753 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553773 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553844 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553867 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-system-cni-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553892 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553912 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553932 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553953 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxsd6\" (UniqueName: \"kubernetes.io/projected/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-kube-api-access-zxsd6\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553975 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.553996 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554016 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554036 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mv9\" (UniqueName: \"kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554062 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-rootfs\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554082 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-os-release\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554093 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554104 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554123 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554144 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-proxy-tls\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554224 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554262 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554282 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-binary-copy\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554304 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554323 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554346 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554373 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554395 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554417 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554437 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554460 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554516 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554553 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554714 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554763 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cnibin\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554792 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554822 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554851 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.554880 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-system-cni-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555151 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555199 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555251 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555582 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555672 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555727 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-rootfs\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555797 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555833 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-binary-copy\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555839 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555809 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-os-release\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555936 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.555987 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.556057 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.556414 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.556454 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.556566 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.556911 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.558160 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-proxy-tls\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.558961 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.566185 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.575124 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxsd6\" (UniqueName: \"kubernetes.io/projected/8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2-kube-api-access-zxsd6\") pod \"multus-additional-cni-plugins-gbqr2\" (UID: \"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\") " pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.575450 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mv9\" (UniqueName: \"kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9\") pod \"ovnkube-node-9qmqs\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.578911 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb8vr\" (UniqueName: \"kubernetes.io/projected/57f7d8bc-8a1d-4b44-be6e-d702c34203c8-kube-api-access-jb8vr\") pod \"machine-config-daemon-x6wwr\" (UID: \"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\") " pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.583254 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.598591 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.624370 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.626900 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.626947 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.626959 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.626976 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.626988 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.644606 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.662824 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.677561 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.680170 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.682747 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerStarted","Data":"e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.682813 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerStarted","Data":"55bc1bd4cb6f1b20ea152f063b261f9801c8b69c055dedb9511ff5a1def9a942"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.683949 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nzqf7" event={"ID":"d7629148-8035-4115-b7f7-89c2868cad0c","Type":"ContainerStarted","Data":"329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.683986 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nzqf7" event={"ID":"d7629148-8035-4115-b7f7-89c2868cad0c","Type":"ContainerStarted","Data":"63c959ed0e363d44d81089eb8ebc0e4176974e7fd741197e19c06f3fd9952d8c"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.684761 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:18 crc kubenswrapper[4899]: W0123 08:52:18.691527 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f9a9cb2_ec5b_4498_ab3e_0061bc62e7b2.slice/crio-07442e5646ec7e8f69329772296bda30db3d6b5c396460949983028fbdc77429 WatchSource:0}: Error finding container 07442e5646ec7e8f69329772296bda30db3d6b5c396460949983028fbdc77429: Status 404 returned error can't find the container with id 07442e5646ec7e8f69329772296bda30db3d6b5c396460949983028fbdc77429 Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.693794 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:52:18 crc kubenswrapper[4899]: E0123 08:52:18.699743 4899 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.699969 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: W0123 08:52:18.707310 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57f7d8bc_8a1d_4b44_be6e_d702c34203c8.slice/crio-a9c4ac45812aa1ae4710c7cb64566fa62f9b7f2876b781f3fe5afbb63e96c4a5 WatchSource:0}: Error finding container a9c4ac45812aa1ae4710c7cb64566fa62f9b7f2876b781f3fe5afbb63e96c4a5: Status 404 returned error can't find the container with id a9c4ac45812aa1ae4710c7cb64566fa62f9b7f2876b781f3fe5afbb63e96c4a5 Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.712589 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.723994 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.729329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.729365 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.729375 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.729390 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.729411 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.738699 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.755114 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.782473 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.799956 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.823189 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.843945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.843978 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.843986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.844000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.844007 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.859617 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.873316 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.884982 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.898612 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.909176 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.917503 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.932472 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.945989 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.946019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.946030 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.946082 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.946094 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:18Z","lastTransitionTime":"2026-01-23T08:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.946843 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.958542 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.970639 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:18 crc kubenswrapper[4899]: I0123 08:52:18.989309 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.048329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.048370 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.048379 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.048403 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.048416 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.150818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.150851 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.150861 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.150876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.150887 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.252865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.252905 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.252913 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.252928 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.252938 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.355389 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.355422 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.355432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.355445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.355454 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.457807 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.458117 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.458125 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.458141 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.458160 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.485517 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 17:52:28.667760364 +0000 UTC Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.514064 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:19 crc kubenswrapper[4899]: E0123 08:52:19.514200 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.514637 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:19 crc kubenswrapper[4899]: E0123 08:52:19.514715 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.514863 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:19 crc kubenswrapper[4899]: E0123 08:52:19.514932 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.560161 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.560191 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.560200 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.560226 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.560235 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.662920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.662947 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.662955 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.662968 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.662977 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.687573 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" exitCode=0 Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.687645 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.687691 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"ee61a426789003e32a0cde8bc9f972379128b2dd20aa8d64f515eb4b4120ed4b"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.689310 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.689339 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.689351 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"a9c4ac45812aa1ae4710c7cb64566fa62f9b7f2876b781f3fe5afbb63e96c4a5"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.691359 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c" exitCode=0 Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.691424 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.691517 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerStarted","Data":"07442e5646ec7e8f69329772296bda30db3d6b5c396460949983028fbdc77429"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.715909 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.730021 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.743262 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.758297 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771742 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771667 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771757 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771917 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.771937 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.783354 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.795731 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.810623 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.820227 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.879460 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.879518 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.879531 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.879552 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.879565 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.881505 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.888618 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-lmzls"] Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.889044 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.891449 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.891778 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.891871 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.892051 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.897809 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.912700 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.935426 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.954579 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.970061 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.980864 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hlq\" (UniqueName: \"kubernetes.io/projected/85294630-22ed-44e4-99f6-ae7fa8658dbe-kube-api-access-q5hlq\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.980924 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85294630-22ed-44e4-99f6-ae7fa8658dbe-host\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.980949 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/85294630-22ed-44e4-99f6-ae7fa8658dbe-serviceca\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.982816 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.983665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.983697 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.983708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.983725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.983735 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:19Z","lastTransitionTime":"2026-01-23T08:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:19 crc kubenswrapper[4899]: I0123 08:52:19.997684 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:19Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.010501 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.021501 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.034000 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.047709 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.060872 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.076324 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.081352 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/85294630-22ed-44e4-99f6-ae7fa8658dbe-serviceca\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.081418 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hlq\" (UniqueName: \"kubernetes.io/projected/85294630-22ed-44e4-99f6-ae7fa8658dbe-kube-api-access-q5hlq\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.081446 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85294630-22ed-44e4-99f6-ae7fa8658dbe-host\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.081508 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85294630-22ed-44e4-99f6-ae7fa8658dbe-host\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.082363 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/85294630-22ed-44e4-99f6-ae7fa8658dbe-serviceca\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.086380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.086411 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.086419 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.086433 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.086442 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.089885 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.100846 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hlq\" (UniqueName: \"kubernetes.io/projected/85294630-22ed-44e4-99f6-ae7fa8658dbe-kube-api-access-q5hlq\") pod \"node-ca-lmzls\" (UID: \"85294630-22ed-44e4-99f6-ae7fa8658dbe\") " pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.104271 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.116566 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.134248 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.194067 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.194098 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.194106 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.194122 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.194132 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.225434 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lmzls" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.296320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.296366 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.296380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.296399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.296411 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.399631 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.399663 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.399674 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.399688 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.399698 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.486163 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 10:51:29.019377381 +0000 UTC Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.502510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.502551 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.502565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.502585 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.502600 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.605942 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.605993 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.606008 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.606028 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.606044 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698484 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698526 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698537 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698546 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698557 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.698565 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.700412 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lmzls" event={"ID":"85294630-22ed-44e4-99f6-ae7fa8658dbe","Type":"ContainerStarted","Data":"61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.700787 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lmzls" event={"ID":"85294630-22ed-44e4-99f6-ae7fa8658dbe","Type":"ContainerStarted","Data":"abc74866c76a0075098f796348bcd067a354568f591033d72559d1a49586e6e6"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.703134 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b" exitCode=0 Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.703180 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.710716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.710756 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.710767 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.710783 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.710794 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.716910 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.736307 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.753026 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.763349 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.781745 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.797420 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.813268 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.813304 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.813314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.813329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.813338 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.816118 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.836491 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.848453 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.863153 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.880524 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.894519 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919333 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919765 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919803 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919815 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919830 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.919843 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:20Z","lastTransitionTime":"2026-01-23T08:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.932058 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.947882 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.961451 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.971480 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.983777 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:20 crc kubenswrapper[4899]: I0123 08:52:20.993956 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:20Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.005708 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.017332 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.022844 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.022880 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.022892 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.022909 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.022921 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.031515 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.042764 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.056865 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.065832 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.080266 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.089381 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.110748 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.125396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.125430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.125438 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.125452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.125461 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.196347 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.196539 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:29.196524425 +0000 UTC m=+36.063494812 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.227864 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.227899 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.227908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.227920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.227929 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.297473 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.297525 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.297551 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.297578 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297673 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297725 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:29.297709999 +0000 UTC m=+36.164680386 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297738 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297794 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297739 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297820 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297884 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297911 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:29.297886904 +0000 UTC m=+36.164857331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297923 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297944 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.297964 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:29.297945645 +0000 UTC m=+36.164916072 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.298021 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:29.297998867 +0000 UTC m=+36.164969274 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.330135 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.330239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.330261 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.330290 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.330313 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.432902 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.432939 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.432949 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.432965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.432976 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.487306 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 14:44:49.956741529 +0000 UTC Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.513800 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.513864 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.513815 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.513931 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.514013 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:21 crc kubenswrapper[4899]: E0123 08:52:21.514173 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.535782 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.535838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.535852 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.535869 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.535881 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.638088 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.638129 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.638141 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.638157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.638171 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.708923 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824" exitCode=0 Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.708978 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.728133 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.740927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.740957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.740966 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.740978 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.740987 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.744404 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.755754 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.770465 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.783984 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.796282 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.808837 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.819819 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.837292 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.842798 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.842829 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.842838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.842852 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.842860 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.847333 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.858487 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.870130 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.881674 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.893160 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:21Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.944541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.944575 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.944584 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.944597 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:21 crc kubenswrapper[4899]: I0123 08:52:21.944609 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:21Z","lastTransitionTime":"2026-01-23T08:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.046864 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.046911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.046922 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.046941 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.046953 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.148884 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.148917 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.148925 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.148937 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.148945 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.251311 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.251362 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.251374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.251391 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.251403 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.353083 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.353116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.353124 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.353138 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.353146 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.455967 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.456013 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.456027 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.456043 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.456054 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.487684 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 21:26:52.624645616 +0000 UTC Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.558328 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.558376 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.558386 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.558399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.558408 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.660530 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.660592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.660605 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.660624 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.660636 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.715978 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25" exitCode=0 Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.716061 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.726135 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.735101 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.749307 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.762467 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.763305 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.763358 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.763368 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.763380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.763389 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.775042 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.789235 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.804348 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.816551 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.829815 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.841848 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.855336 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865202 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865253 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865284 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865308 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.865924 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.882309 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.906603 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.922276 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:22Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.967814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.967851 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.967861 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.967878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:22 crc kubenswrapper[4899]: I0123 08:52:22.967888 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:22Z","lastTransitionTime":"2026-01-23T08:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.070413 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.070720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.070730 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.070744 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.070754 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.179490 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.179531 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.179548 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.179565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.179576 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.281034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.281069 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.281077 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.281101 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.281109 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.383655 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.383699 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.383714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.383734 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.383749 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.391799 4899 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.488709 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 08:56:12.598576004 +0000 UTC Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.497680 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.497749 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.497763 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.497786 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.497801 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.513846 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.513845 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.514031 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:23 crc kubenswrapper[4899]: E0123 08:52:23.513986 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:23 crc kubenswrapper[4899]: E0123 08:52:23.514084 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:23 crc kubenswrapper[4899]: E0123 08:52:23.514134 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.531590 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.541706 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.561097 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.574098 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.585318 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.596040 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.603810 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.603835 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.603858 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.603871 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.603880 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.607317 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.618154 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.629778 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.639933 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.654887 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.670010 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.688970 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.700583 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.705608 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.705661 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.705670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.705684 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.705692 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.734996 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354" exitCode=0 Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.735055 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.748797 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.758277 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.767236 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.779044 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.788584 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.799200 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809522 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809579 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809597 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809620 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809637 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.809872 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.820246 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.833622 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.843829 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.857095 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.868499 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.889043 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.900352 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.912253 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.912304 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.912318 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.912336 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:23 crc kubenswrapper[4899]: I0123 08:52:23.912348 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:23Z","lastTransitionTime":"2026-01-23T08:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.014844 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.014908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.014918 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.014932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.014943 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.118569 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.118650 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.118668 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.118697 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.118715 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.221821 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.221889 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.221908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.221933 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.221955 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.324720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.324769 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.324785 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.324804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.324820 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.427754 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.427818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.427839 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.427863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.427880 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.489293 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 21:21:56.763029 +0000 UTC Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.530420 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.530488 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.530510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.530540 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.530568 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.632635 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.632977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.632987 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.633005 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.633017 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.735559 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.735618 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.735640 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.735662 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.735678 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.746872 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.747808 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.747864 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.752041 4899 generic.go:334] "Generic (PLEG): container finished" podID="8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2" containerID="caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c" exitCode=0 Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.752066 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerDied","Data":"caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.763710 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.774742 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.775049 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.780199 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.795442 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.805122 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.818668 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.833091 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.838129 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.838238 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.838270 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.838339 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.838363 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.847001 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.857960 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.871879 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.883445 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.894432 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.907754 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.924638 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.935409 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.940456 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.940503 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.940519 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.940538 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.940554 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:24Z","lastTransitionTime":"2026-01-23T08:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.950534 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.963051 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.974093 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.987382 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:24 crc kubenswrapper[4899]: I0123 08:52:24.998624 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:24Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.012587 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.025330 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.036369 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.042792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.042830 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.042838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.042851 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.042859 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.052698 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.062747 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.073708 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.088505 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.104357 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.119072 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.144609 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.144651 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.144663 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.144680 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.144693 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.247315 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.247367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.247384 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.247405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.247418 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.349838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.349895 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.349918 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.349965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.349990 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.453548 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.453615 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.453632 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.453656 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.453674 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.490320 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 16:46:18.369018905 +0000 UTC Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.514176 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:25 crc kubenswrapper[4899]: E0123 08:52:25.514408 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.514231 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.514528 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:25 crc kubenswrapper[4899]: E0123 08:52:25.514682 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:25 crc kubenswrapper[4899]: E0123 08:52:25.514844 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.556601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.556653 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.556670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.556691 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.556706 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.659866 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.659922 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.659933 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.659953 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.659970 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.760377 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" event={"ID":"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2","Type":"ContainerStarted","Data":"71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.760962 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.761900 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.761960 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.761978 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.762005 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.762027 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.775289 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.787695 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.806766 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.814760 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.826897 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.845471 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.857515 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.864314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.864342 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.864351 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.864363 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.864370 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.870746 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.882966 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.894811 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.908706 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.922714 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.936016 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.948522 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:25Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.966266 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.966303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.966311 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.966325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:25 crc kubenswrapper[4899]: I0123 08:52:25.966334 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:25Z","lastTransitionTime":"2026-01-23T08:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.068594 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.068820 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.068879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.068977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.069038 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.171923 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.171963 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.171974 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.171988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.171996 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.274130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.274174 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.274183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.274197 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.274222 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.376429 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.376502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.376516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.376533 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.376545 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.478590 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.478628 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.478638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.478656 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.478667 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.491076 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 23:11:51.643685101 +0000 UTC Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.580533 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.580571 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.580580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.580598 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.580606 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.683948 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.683992 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.684004 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.684031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.684063 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.763653 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.786166 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.786223 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.786235 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.786252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.786264 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.830703 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.830738 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.830748 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.830761 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.830772 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.847922 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:26Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.851641 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.851703 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.851716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.851753 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.851765 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.874714 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:26Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.882108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.882160 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.882175 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.882196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.882236 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.903250 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:26Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.906275 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.906304 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.906317 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.906331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.906342 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.921775 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:26Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.925046 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.925167 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.925283 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.925388 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.925482 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.943509 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:26Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:26 crc kubenswrapper[4899]: E0123 08:52:26.943621 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.945170 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.945203 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.945226 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.945239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:26 crc kubenswrapper[4899]: I0123 08:52:26.945250 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:26Z","lastTransitionTime":"2026-01-23T08:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.047977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.048231 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.048294 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.048392 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.048454 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.151769 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.151812 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.151857 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.151908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.151923 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.254409 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.254452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.254462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.254477 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.254488 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.356291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.356649 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.356662 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.356678 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.356689 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.459581 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.459614 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.459623 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.459642 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.459653 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.492188 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 05:10:32.194867357 +0000 UTC Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.514036 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.514128 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:27 crc kubenswrapper[4899]: E0123 08:52:27.514309 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.514343 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:27 crc kubenswrapper[4899]: E0123 08:52:27.514433 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:27 crc kubenswrapper[4899]: E0123 08:52:27.514562 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.562372 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.562411 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.562423 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.562438 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.562450 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.667145 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.667183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.667195 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.667225 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.667235 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.769291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.769537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.769555 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.769588 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.769605 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.771329 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/0.log" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.774677 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376" exitCode=1 Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.774712 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.775349 4899 scope.go:117] "RemoveContainer" containerID="affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.790881 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.805382 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.818249 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.830114 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.846164 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.864821 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.872236 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.872292 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.872307 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.872324 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.872337 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.875963 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.900550 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.912332 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.934625 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.947992 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.962394 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.974408 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.980031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.980095 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.980109 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.980130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.980147 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:27Z","lastTransitionTime":"2026-01-23T08:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:27 crc kubenswrapper[4899]: I0123 08:52:27.988539 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:27Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.084363 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.084432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.084442 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.084455 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.084464 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.187005 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.187038 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.187048 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.187061 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.187070 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.289780 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.289822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.289834 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.289849 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.289862 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.392308 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.392342 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.392351 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.392365 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.392374 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.492798 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 00:36:09.509337424 +0000 UTC Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.494136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.494167 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.494179 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.494195 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.494210 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.596120 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.596162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.596178 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.596197 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.596256 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.699964 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.700025 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.700047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.700076 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.700096 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.780453 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/0.log" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.783392 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.783498 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.801036 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.802127 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.802151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.802158 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.802170 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.802181 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.817124 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.833655 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.850529 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.864626 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.884627 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.894638 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.904489 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.904532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.904543 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.904561 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.904575 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:28Z","lastTransitionTime":"2026-01-23T08:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.910363 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.924800 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.937964 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.956909 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.970558 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.985483 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:28 crc kubenswrapper[4899]: I0123 08:52:28.997767 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.006506 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.006554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.006567 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.006586 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.006599 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.108603 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.108645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.108658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.108673 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.108684 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.197849 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.198038 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:45.198000005 +0000 UTC m=+52.064970432 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.211863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.211933 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.211951 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.211972 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.211989 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.298491 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.298551 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.298598 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.298640 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298739 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298741 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298782 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298801 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298809 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:45.298788648 +0000 UTC m=+52.165759045 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298813 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298830 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298944 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298961 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.298847 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:45.298833639 +0000 UTC m=+52.165804046 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.299024 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:45.299003714 +0000 UTC m=+52.165974181 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.299039 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:52:45.299031434 +0000 UTC m=+52.166001921 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.314728 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.314765 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.314777 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.314792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.314804 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.418370 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.418452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.418471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.418503 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.418522 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.493651 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 21:44:58.363843889 +0000 UTC Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.514785 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.514843 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.515036 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.515019 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.515174 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.515275 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.520602 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.520631 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.520640 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.520652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.520661 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.623520 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.623558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.623568 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.623585 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.623595 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.725032 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.725088 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.725111 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.725247 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.725321 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.788930 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/1.log" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.789621 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/0.log" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.793327 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091" exitCode=1 Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.793372 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.793409 4899 scope.go:117] "RemoveContainer" containerID="affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.794715 4899 scope.go:117] "RemoveContainer" containerID="6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091" Jan 23 08:52:29 crc kubenswrapper[4899]: E0123 08:52:29.795035 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.809567 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.824515 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.827190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.827250 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.827258 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.827272 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.827282 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.842128 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.851904 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.864475 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.875635 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.886900 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.899391 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.911144 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.923687 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.929814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.929863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.929878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.929899 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.929914 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:29Z","lastTransitionTime":"2026-01-23T08:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.932532 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.945803 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.961748 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:29 crc kubenswrapper[4899]: I0123 08:52:29.970528 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:29Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.037888 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.037965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.037980 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.038000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.038014 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.141249 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.141306 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.141319 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.141340 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.141352 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.244535 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.244621 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.244632 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.244647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.244660 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.347141 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.347184 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.347193 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.347229 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.347238 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.449765 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.449804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.449814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.449828 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.449838 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.494804 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 09:34:00.129672541 +0000 UTC Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.515335 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.553927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.553957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.553965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.553979 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.553987 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.563142 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn"] Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.563561 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.566123 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.569357 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.581177 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.595943 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.608489 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.612292 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.612400 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flg5c\" (UniqueName: \"kubernetes.io/projected/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-kube-api-access-flg5c\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.612457 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.612480 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.618917 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.632775 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.644516 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.656367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.656425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.656437 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.656457 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.656470 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.666191 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.680059 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.695165 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.710074 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.712988 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.713033 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flg5c\" (UniqueName: \"kubernetes.io/projected/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-kube-api-access-flg5c\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.713064 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.713461 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.713627 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.714018 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.719950 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.726302 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.731120 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flg5c\" (UniqueName: \"kubernetes.io/projected/28c6ba1e-01d3-4b6f-b539-5279a6cbfed0-kube-api-access-flg5c\") pod \"ovnkube-control-plane-749d76644c-khwdn\" (UID: \"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.738567 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.752559 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.759965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.760026 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.760038 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.760060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.760072 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.764456 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.775143 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:30Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.798286 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/1.log" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.862018 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.862050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.862057 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.862075 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.862085 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.886681 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" Jan 23 08:52:30 crc kubenswrapper[4899]: W0123 08:52:30.897811 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28c6ba1e_01d3_4b6f_b539_5279a6cbfed0.slice/crio-ee9b35a3edd43ac116cc829c33f92b2ce431d6c56b7cac6ac7d9765d7b877d44 WatchSource:0}: Error finding container ee9b35a3edd43ac116cc829c33f92b2ce431d6c56b7cac6ac7d9765d7b877d44: Status 404 returned error can't find the container with id ee9b35a3edd43ac116cc829c33f92b2ce431d6c56b7cac6ac7d9765d7b877d44 Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.968865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.968901 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.968910 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.968923 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:30 crc kubenswrapper[4899]: I0123 08:52:30.968938 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:30Z","lastTransitionTime":"2026-01-23T08:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.071383 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.071428 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.071440 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.071456 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.071468 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.174159 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.174200 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.174221 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.174240 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.174251 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.277171 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.277263 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.277282 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.277303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.277320 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.379277 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.379316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.379325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.379341 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.379351 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.481398 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.481462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.481484 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.481511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.481531 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.495988 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 19:13:45.137595921 +0000 UTC Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.514526 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.514544 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:31 crc kubenswrapper[4899]: E0123 08:52:31.514709 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.514556 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:31 crc kubenswrapper[4899]: E0123 08:52:31.514872 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:31 crc kubenswrapper[4899]: E0123 08:52:31.515065 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.584255 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.584288 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.584299 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.584315 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.584326 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.687335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.687400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.687422 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.687452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.687473 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.790079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.790116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.790125 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.790140 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.790153 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.806760 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" event={"ID":"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0","Type":"ContainerStarted","Data":"504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.806839 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" event={"ID":"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0","Type":"ContainerStarted","Data":"aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.806852 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" event={"ID":"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0","Type":"ContainerStarted","Data":"ee9b35a3edd43ac116cc829c33f92b2ce431d6c56b7cac6ac7d9765d7b877d44"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.808889 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.810726 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.811107 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.827432 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.842570 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.854995 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.870034 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.880888 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.892497 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.892532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.892541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.892554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.892563 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.895029 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.906412 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.918593 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.930672 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.952905 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.966651 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.977562 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.989175 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:31Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.995117 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.995155 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.995165 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.995180 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:31 crc kubenswrapper[4899]: I0123 08:52:31.995189 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:31Z","lastTransitionTime":"2026-01-23T08:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.004648 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.018300 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.031529 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.040421 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kxcxw"] Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.040861 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: E0123 08:52:32.040923 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.045962 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.055942 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.069424 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.081096 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.090304 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.097376 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.097412 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.097421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.097433 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.097442 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.103437 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.111826 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.121704 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.126012 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7l67\" (UniqueName: \"kubernetes.io/projected/bb844e06-c862-406b-8ecb-36b682d2d58b-kube-api-access-v7l67\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.126093 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.138579 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.148966 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.165579 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.180886 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.195112 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.200264 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.200331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.200354 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.200376 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.200392 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.214881 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.226561 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7l67\" (UniqueName: \"kubernetes.io/projected/bb844e06-c862-406b-8ecb-36b682d2d58b-kube-api-access-v7l67\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.226604 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: E0123 08:52:32.226698 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:32 crc kubenswrapper[4899]: E0123 08:52:32.226749 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:32.726734239 +0000 UTC m=+39.593704636 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.246986 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.253709 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7l67\" (UniqueName: \"kubernetes.io/projected/bb844e06-c862-406b-8ecb-36b682d2d58b-kube-api-access-v7l67\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.258833 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.271410 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.286236 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.298733 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.302337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.302372 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.302383 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.302399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.302412 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.309921 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.320165 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.329186 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.338611 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.351568 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.362223 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.379578 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.389418 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.399583 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.404386 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.404426 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.404435 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.404449 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.404458 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.414271 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.427950 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:32Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.496644 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 23:37:21.071883259 +0000 UTC Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.506472 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.506510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.506519 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.506532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.506541 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.609286 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.609348 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.609358 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.609372 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.609381 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.712426 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.712465 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.712479 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.712493 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.712503 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.731457 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:32 crc kubenswrapper[4899]: E0123 08:52:32.731610 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:32 crc kubenswrapper[4899]: E0123 08:52:32.731687 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:33.731668792 +0000 UTC m=+40.598639179 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.813988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.814050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.814060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.814074 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.814086 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.916654 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.916695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.916705 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.916723 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:32 crc kubenswrapper[4899]: I0123 08:52:32.916735 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:32Z","lastTransitionTime":"2026-01-23T08:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.038040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.038086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.038097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.038113 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.038126 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.140493 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.140526 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.140534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.140547 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.140556 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.242818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.242860 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.242870 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.242886 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.242898 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.345549 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.345590 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.345601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.345614 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.345624 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.448029 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.448121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.448136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.448170 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.448182 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.497431 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 09:50:30.892539967 +0000 UTC Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.514261 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.514396 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.514506 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.514401 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.514599 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.514671 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.514818 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.514881 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.541541 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://affaafcee9ecf6ffa70e01cbab8253f7cf5b6adab985d05d8e42712319d48376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:27Z\\\",\\\"message\\\":\\\"Stopping watch factory\\\\nI0123 08:52:27.685568 6229 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.685646 6229 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:27.685659 6229 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:27.685648 6229 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:27.685670 6229 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0123 08:52:27.685688 6229 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685671 6229 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:27.685660 6229 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:52:27.685841 6229 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0123 08:52:27.686171 6229 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.550713 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.550772 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.550794 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.550816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.550833 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.552872 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.562390 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.572716 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.584544 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.596806 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.608838 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.620398 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.630808 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.644199 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.653573 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.653622 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.653633 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.653647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.653657 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.659433 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.672528 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.687069 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.699615 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.714968 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.727456 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:33Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.744886 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.745014 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:33 crc kubenswrapper[4899]: E0123 08:52:33.745104 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:35.745083501 +0000 UTC m=+42.612053968 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.755527 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.755596 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.755604 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.755839 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.755851 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.857988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.858034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.858046 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.858063 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.858076 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.960432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.960504 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.960521 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.960543 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:33 crc kubenswrapper[4899]: I0123 08:52:33.960557 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:33Z","lastTransitionTime":"2026-01-23T08:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.063656 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.064138 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.064407 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.064641 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.064846 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.168576 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.168620 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.168633 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.168651 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.168664 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.271936 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.271978 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.271986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.272001 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.272010 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.374949 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.375810 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.375844 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.375874 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.375896 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.478170 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.478240 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.478256 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.478270 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.478621 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.497514 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 17:01:16.585791698 +0000 UTC Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.580987 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.581025 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.581043 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.581059 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.581071 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.683732 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.683787 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.683799 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.683812 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.683823 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.786571 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.786612 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.786626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.786643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.786656 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.889506 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.889599 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.889614 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.889632 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.889645 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.992617 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.992657 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.992666 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.992680 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:34 crc kubenswrapper[4899]: I0123 08:52:34.992694 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:34Z","lastTransitionTime":"2026-01-23T08:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.094676 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.094967 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.095037 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.095105 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.095161 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.197992 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.198044 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.198060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.198081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.198094 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.300947 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.301006 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.301024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.301046 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.301062 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.403895 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.403976 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.404003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.404038 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.404061 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.498596 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 07:04:30.784864672 +0000 UTC Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.507435 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.507529 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.507541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.507560 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.507573 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.513948 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.513985 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.514069 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.514116 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.514175 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.514326 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.514542 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.514593 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.609784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.609831 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.609841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.609865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.609878 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.712603 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.712670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.712683 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.712701 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.712713 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.762844 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.763106 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:35 crc kubenswrapper[4899]: E0123 08:52:35.763281 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:39.763250141 +0000 UTC m=+46.630220568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.814992 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.815046 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.815059 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.815076 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.815089 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.918181 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.918366 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.918395 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.918477 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:35 crc kubenswrapper[4899]: I0123 08:52:35.918505 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:35Z","lastTransitionTime":"2026-01-23T08:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.021647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.021715 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.021733 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.021755 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.021770 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.126530 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.126730 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.126792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.126822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.126841 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.232652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.232714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.232727 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.232756 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.232769 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.335926 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.335964 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.335973 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.335986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.335995 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.438896 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.438975 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.438987 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.439003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.439016 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.499769 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 20:39:51.626114356 +0000 UTC Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.541150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.541188 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.541196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.541224 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.541234 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.644152 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.644199 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.644227 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.644241 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.644250 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.746616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.746879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.746898 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.746916 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.746927 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.849831 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.849884 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.849895 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.849911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.849922 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.951925 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.952151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.952259 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.952359 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:36 crc kubenswrapper[4899]: I0123 08:52:36.952433 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:36Z","lastTransitionTime":"2026-01-23T08:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.054470 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.054692 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.054790 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.054855 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.054910 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.098318 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.098359 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.098369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.098389 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.098408 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.110790 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:37Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.114050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.114089 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.114102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.114115 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.114123 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.124626 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:37Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.127856 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.127896 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.127908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.127926 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.127938 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.138910 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:37Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.142495 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.142531 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.142541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.142554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.142564 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.154486 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:37Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.157541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.157573 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.157584 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.157602 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.157613 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.168763 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:37Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.168875 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.170291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.170319 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.170333 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.170350 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.170362 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.272223 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.272475 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.272544 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.272637 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.272713 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.374936 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.374975 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.374984 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.374998 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.375006 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.477851 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.477917 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.477935 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.477958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.477975 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.500880 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:09:13.541307198 +0000 UTC Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.513891 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.513925 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.514042 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.514071 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.514176 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.514306 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.514420 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:37 crc kubenswrapper[4899]: E0123 08:52:37.514534 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.581579 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.581619 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.581645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.581666 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.581684 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.683984 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.684099 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.684115 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.684144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.684347 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.786655 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.786701 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.786712 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.786730 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.786742 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.889115 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.889166 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.889190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.889231 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.889251 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.991964 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.992002 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.992012 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.992026 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:37 crc kubenswrapper[4899]: I0123 08:52:37.992038 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:37Z","lastTransitionTime":"2026-01-23T08:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.095141 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.095427 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.095549 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.095665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.095742 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.198330 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.198369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.198413 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.198427 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.198441 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.300572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.300626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.300642 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.300660 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.300671 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.402867 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.402917 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.402928 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.402945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.402959 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.501024 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 07:19:51.881668334 +0000 UTC Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.504899 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.504946 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.504958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.504975 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.504989 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.607959 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.608016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.608034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.608056 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.608073 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.710132 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.710178 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.710192 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.710212 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.710239 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.813101 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.813145 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.813157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.813173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.813185 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.915828 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.915876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.915887 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.915902 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:38 crc kubenswrapper[4899]: I0123 08:52:38.915914 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:38Z","lastTransitionTime":"2026-01-23T08:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.018499 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.018539 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.018551 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.018570 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.018583 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.120940 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.120979 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.120991 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.121006 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.121022 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.224716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.225232 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.225445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.225647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.225790 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.329164 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.329234 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.329245 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.329259 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.329270 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.432194 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.432526 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.432704 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.432938 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.433182 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.501292 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 06:30:21.968623245 +0000 UTC Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.513854 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.514329 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.514414 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.514385 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.514489 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.514529 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.514645 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.514881 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.536367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.536439 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.536462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.536488 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.536506 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.639649 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.639688 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.639698 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.639713 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.639726 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.741607 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.741712 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.741731 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.741799 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.741816 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.808967 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.809311 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:39 crc kubenswrapper[4899]: E0123 08:52:39.809448 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:52:47.809417401 +0000 UTC m=+54.676387828 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.845177 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.845338 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.845369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.845400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.845422 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.947598 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.948107 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.948198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.948301 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:39 crc kubenswrapper[4899]: I0123 08:52:39.948381 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:39Z","lastTransitionTime":"2026-01-23T08:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.050763 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.050972 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.050982 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.051000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.051013 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.153547 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.153594 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.153605 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.153623 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.153637 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.256825 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.256876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.256887 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.256905 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.256917 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.358952 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.358999 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.359010 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.359028 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.359039 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.460855 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.460995 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.461016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.461040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.461057 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.501974 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 17:18:48.910901577 +0000 UTC Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.514148 4899 scope.go:117] "RemoveContainer" containerID="6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.537968 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.548946 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.559982 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.562907 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.562940 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.562950 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.562964 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.562975 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.574015 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.588037 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.600990 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.610771 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.623527 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.637439 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.650393 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.663352 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.665513 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.665549 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.665563 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.665586 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.665607 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.675955 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.686988 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.698188 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.709871 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.719995 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.769275 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.769341 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.769364 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.769394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.769423 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.841482 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/1.log" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.846644 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.847143 4899 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.862554 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.872557 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.872599 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.872609 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.872629 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.872640 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.879909 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.891090 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.902987 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.913819 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.926114 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.942765 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.953442 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.967460 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.974989 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.975025 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.975038 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.975055 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.975067 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:40Z","lastTransitionTime":"2026-01-23T08:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:40 crc kubenswrapper[4899]: I0123 08:52:40.982599 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:40Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.005730 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.024553 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.050898 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.068455 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.077421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.077454 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.077463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.077476 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.077487 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.085106 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.097185 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.179536 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.179839 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.179855 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.179872 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.179882 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.282474 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.282513 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.282523 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.282537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.282547 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.385818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.385881 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.385897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.385921 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.385935 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.488958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.489019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.489036 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.489061 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.489078 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.502738 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 18:44:23.442914315 +0000 UTC Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.514537 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.514533 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:41 crc kubenswrapper[4899]: E0123 08:52:41.514712 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.514783 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.514804 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:41 crc kubenswrapper[4899]: E0123 08:52:41.514920 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:41 crc kubenswrapper[4899]: E0123 08:52:41.515077 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:41 crc kubenswrapper[4899]: E0123 08:52:41.515152 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.591176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.591239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.591255 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.591273 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.591284 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.694591 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.694661 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.694683 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.694711 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.694734 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.798133 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.798254 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.798283 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.798313 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.798331 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.900791 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.900869 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.900886 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.900908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:41 crc kubenswrapper[4899]: I0123 08:52:41.900923 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:41Z","lastTransitionTime":"2026-01-23T08:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.003912 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.003961 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.003977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.004026 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.004038 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.106897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.107136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.107158 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.107189 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.107254 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.209735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.209782 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.209793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.209808 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.209820 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.311297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.311334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.311351 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.311369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.311392 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.413183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.413248 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.413256 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.413269 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.413278 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.503663 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:37:36.669374396 +0000 UTC Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.515588 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.515640 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.515655 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.515676 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.515692 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.618848 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.619050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.619081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.619110 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.619133 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.723362 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.723487 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.723511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.723542 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.723565 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.826011 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.826069 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.826084 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.826098 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.826108 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.853511 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/2.log" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.853979 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/1.log" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.857066 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" exitCode=1 Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.857110 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.857148 4899 scope.go:117] "RemoveContainer" containerID="6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.858999 4899 scope.go:117] "RemoveContainer" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" Jan 23 08:52:42 crc kubenswrapper[4899]: E0123 08:52:42.859322 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.880116 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.890386 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.912470 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.925977 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.928565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.928596 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.928634 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.928655 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.928666 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:42Z","lastTransitionTime":"2026-01-23T08:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.942651 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.955416 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.968450 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.979823 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:42 crc kubenswrapper[4899]: I0123 08:52:42.990984 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:42Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.003759 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.018484 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.030547 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.030571 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.030580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.030593 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.030601 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.036429 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.048438 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.071146 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.082590 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.100043 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.132396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.132461 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.132484 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.132515 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.132537 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.235413 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.235479 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.235502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.235531 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.235552 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.338070 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.338257 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.338285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.338314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.338336 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.441007 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.441054 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.441068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.441085 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.441098 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.504338 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 07:46:17.30989063 +0000 UTC Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.513873 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.513971 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:43 crc kubenswrapper[4899]: E0123 08:52:43.514017 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.514071 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.514246 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:43 crc kubenswrapper[4899]: E0123 08:52:43.514203 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:43 crc kubenswrapper[4899]: E0123 08:52:43.514393 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:43 crc kubenswrapper[4899]: E0123 08:52:43.514462 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.535081 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.544129 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.544173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.544184 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.544203 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.544247 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.550643 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.570395 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.587571 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.601949 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.611870 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.630770 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.645548 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.646572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.646661 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.646679 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.646704 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.646722 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.658675 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.678186 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.690803 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.703668 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.718726 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.731920 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.748401 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.750060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.750124 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.750137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.750157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.750173 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.763357 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:43Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.851911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.851954 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.852029 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.852071 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.852085 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.862068 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/2.log" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.954243 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.954289 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.954299 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.954316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:43 crc kubenswrapper[4899]: I0123 08:52:43.954327 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:43Z","lastTransitionTime":"2026-01-23T08:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.057388 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.057425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.057436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.057451 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.057461 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.160467 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.160505 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.160515 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.160531 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.160547 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.263017 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.263069 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.263080 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.263094 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.263105 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.365814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.365863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.365876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.365893 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.365905 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.469099 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.469158 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.469177 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.469239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.469259 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.505574 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 06:27:28.779849154 +0000 UTC Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.571791 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.571861 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.571881 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.571910 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.571934 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.674717 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.674773 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.674790 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.674813 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.674830 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.777879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.777914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.777924 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.777939 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.777950 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.879718 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.879758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.879768 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.879786 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.879796 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.982970 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.983034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.983051 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.983079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:44 crc kubenswrapper[4899]: I0123 08:52:44.983100 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:44Z","lastTransitionTime":"2026-01-23T08:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.085285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.085355 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.085370 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.085392 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.085407 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.187927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.188003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.188037 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.188055 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.188066 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.275533 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.275691 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:53:17.275659755 +0000 UTC m=+84.142630182 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.290930 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.290981 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.290999 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.291022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.291044 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.377459 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.377531 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.377572 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.377649 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377648 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377705 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377682 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377807 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377818 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377839 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:53:17.377800584 +0000 UTC m=+84.244771011 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377869 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:53:17.377855086 +0000 UTC m=+84.244825513 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377894 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:53:17.377882167 +0000 UTC m=+84.244852594 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.377920 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.378155 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.378176 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.378307 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:53:17.378280817 +0000 UTC m=+84.245251214 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.393829 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.393906 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.393928 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.393959 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.393981 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.479781 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.490748 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.499436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.499466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.499474 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.499486 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.499495 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.505525 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.505729 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 14:17:33.088382123 +0000 UTC Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.514038 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.514179 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.514241 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.514308 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.514340 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.514389 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.514452 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:45 crc kubenswrapper[4899]: E0123 08:52:45.514633 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.519031 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.532086 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.545840 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.561553 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.574190 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.584753 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.594537 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.601739 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.601779 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.601789 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.601805 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.601816 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.605186 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.619119 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.628626 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.636755 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.648409 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.658409 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.668394 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.688577 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.704130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.704164 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.704176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.704194 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.704207 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.806746 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.806914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.806940 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.806965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.806983 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.910342 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.910403 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.910425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.910453 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.910476 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:45Z","lastTransitionTime":"2026-01-23T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.954348 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:52:45 crc kubenswrapper[4899]: I0123 08:52:45.982995 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db37e5716c192522917f36dd73b8ef8104db7f23012ba6844fb68e480ebd091\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:29Z\\\",\\\"message\\\":\\\" Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:52:28.516978 6361 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:52:28.516993 6361 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:52:28.517004 6361 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:52:28.517011 6361 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:52:28.517017 6361 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:52:28.517023 6361 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:52:28.517041 6361 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:52:28.517060 6361 factory.go:656] Stopping watch factory\\\\nI0123 08:52:28.517072 6361 ovnkube.go:599] Stopped ovnkube\\\\nI0123 08:52:28.517082 6361 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:52:28.517096 6361 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:52:28.517098 6361 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:52:28.517101 6361 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0123 08:52:28.517107 6361 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0123 08:52:28.517168 6361 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.000586 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:45Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.013057 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.013119 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.013130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.013148 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.013159 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.018759 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.038804 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.058074 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.072853 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.086879 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.101244 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.114732 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.115676 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.115708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.115719 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.115733 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.115743 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.127847 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.140915 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.154309 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.166863 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.179203 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.188145 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.200760 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.209782 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:46Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.218520 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.218558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.218569 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.218583 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.218593 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.320931 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.320966 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.320984 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.321000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.321010 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.424473 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.424539 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.424557 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.424578 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.424594 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.505807 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 10:23:07.32847003 +0000 UTC Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.527276 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.527320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.527337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.527359 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.527376 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.630639 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.630767 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.630791 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.630819 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.630842 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.733670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.733734 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.733751 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.733773 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.733790 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.836847 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.836888 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.836897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.836912 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.836922 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.939975 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.940052 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.940074 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.940103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:46 crc kubenswrapper[4899]: I0123 08:52:46.940128 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:46Z","lastTransitionTime":"2026-01-23T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.043953 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.044040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.044058 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.044086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.044099 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.044638 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.045493 4899 scope.go:117] "RemoveContainer" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.045665 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.063472 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.075706 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.091711 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.104285 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.124689 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.140532 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.147750 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.147818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.147835 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.147857 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.147874 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.161742 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.177730 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.192330 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.207327 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.222626 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.240454 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.250566 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.251113 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.251183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.251233 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.251263 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.251282 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.262132 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.276675 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.289463 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.302137 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.353624 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.353668 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.353676 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.353692 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.353702 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.456737 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.456776 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.456785 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.456801 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.456810 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.499554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.499599 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.499616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.499637 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.499649 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.506712 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 02:53:35.948441581 +0000 UTC Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.514074 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.514080 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.514144 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.514147 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.514262 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.514386 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.514543 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.514621 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.515824 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.524391 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.524451 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.524471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.524497 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.524516 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.541082 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.545360 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.545393 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.545403 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.545416 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.545426 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.558034 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.562528 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.562567 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.562578 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.562592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.562603 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.602843 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.608362 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.608434 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.608457 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.608485 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.608506 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.634951 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:47Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.635194 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.637121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.637290 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.637397 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.637484 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.637557 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.740016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.740058 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.740067 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.740081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.740092 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.842252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.842314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.842325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.842340 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.842349 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.908865 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.909049 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:47 crc kubenswrapper[4899]: E0123 08:52:47.909134 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:53:03.909116775 +0000 UTC m=+70.776087162 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.945093 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.945180 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.945191 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.945237 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:47 crc kubenswrapper[4899]: I0123 08:52:47.945249 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:47Z","lastTransitionTime":"2026-01-23T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.047026 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.047076 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.047087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.047103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.047114 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.150003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.150059 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.150075 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.150097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.150113 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.252148 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.252195 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.252230 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.252246 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.252259 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.355325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.355377 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.355391 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.355415 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.355432 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.458048 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.458125 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.458149 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.458179 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.458200 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.507339 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 20:26:05.80433093 +0000 UTC Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.560720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.560765 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.560783 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.560809 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.560825 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.663495 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.663555 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.663572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.663593 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.663610 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.765962 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.766004 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.766016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.766033 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.766045 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.868795 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.868847 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.868858 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.868875 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.868886 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.972752 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.972880 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.972903 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.972932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:48 crc kubenswrapper[4899]: I0123 08:52:48.972953 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:48Z","lastTransitionTime":"2026-01-23T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.076679 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.076753 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.076774 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.076802 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.076823 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.179708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.179816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.179840 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.179914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.179939 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.283297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.283407 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.283468 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.283500 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.283522 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.386840 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.386908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.386930 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.386958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.386980 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.492916 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.493007 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.493032 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.493064 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.493094 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.508013 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 11:53:21.411778216 +0000 UTC Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.514516 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.514571 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.514623 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:49 crc kubenswrapper[4899]: E0123 08:52:49.514663 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:49 crc kubenswrapper[4899]: E0123 08:52:49.514872 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:49 crc kubenswrapper[4899]: E0123 08:52:49.514945 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.515141 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:49 crc kubenswrapper[4899]: E0123 08:52:49.515462 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.596798 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.596854 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.596866 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.596882 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.596894 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.699770 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.699829 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.699840 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.699856 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.699865 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.802102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.802137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.802144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.802156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.802166 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.904779 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.904814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.904822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.904834 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:49 crc kubenswrapper[4899]: I0123 08:52:49.904842 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:49Z","lastTransitionTime":"2026-01-23T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.007180 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.007575 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.007594 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.007618 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.007633 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.110176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.110299 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.110364 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.110390 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.110409 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.212482 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.212516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.212524 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.212551 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.212560 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.315638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.315714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.315732 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.315755 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.315772 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.418709 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.418776 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.418797 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.418828 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.418854 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.509048 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 13:56:16.384152796 +0000 UTC Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.520954 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.520986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.520996 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.521007 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.521016 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.623394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.623474 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.623486 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.623510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.623542 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.726097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.726151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.726173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.726194 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.726230 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.829148 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.829198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.829264 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.829289 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.829299 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.931414 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.931463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.931488 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.931508 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:50 crc kubenswrapper[4899]: I0123 08:52:50.931519 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:50Z","lastTransitionTime":"2026-01-23T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.033723 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.033798 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.033813 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.033839 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.033857 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.136134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.136200 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.136246 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.136270 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.136285 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.239302 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.239361 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.239373 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.239397 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.239412 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.341664 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.341704 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.341712 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.341725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.341734 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.444342 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.444748 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.444915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.445091 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.445276 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.509795 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 15:36:39.069218774 +0000 UTC Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.514313 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:51 crc kubenswrapper[4899]: E0123 08:52:51.514519 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.514325 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:51 crc kubenswrapper[4899]: E0123 08:52:51.514780 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.514861 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:51 crc kubenswrapper[4899]: E0123 08:52:51.515338 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.515578 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:51 crc kubenswrapper[4899]: E0123 08:52:51.515725 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.547865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.547915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.547932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.547955 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.547971 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.650380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.650426 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.650436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.650451 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.650460 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.754958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.755134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.755252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.755302 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.755386 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.858222 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.858268 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.858277 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.858291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.858304 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.960878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.960923 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.960934 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.960950 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:51 crc kubenswrapper[4899]: I0123 08:52:51.960962 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:51Z","lastTransitionTime":"2026-01-23T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.064316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.064367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.064379 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.064398 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.064415 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.167039 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.167086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.167097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.167116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.167128 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.270667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.270710 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.270719 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.270734 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.270745 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.373483 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.373534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.373552 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.373569 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.373580 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.476956 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.477008 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.477018 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.477032 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.477041 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.510772 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 10:33:19.157102964 +0000 UTC Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.579418 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.579463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.579472 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.579485 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.579494 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.681335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.681377 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.681385 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.681399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.681410 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.783926 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.783976 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.783988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.784005 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.784021 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.887040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.887079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.887089 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.887103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.887114 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.989493 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.989535 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.989543 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.989558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:52 crc kubenswrapper[4899]: I0123 08:52:52.989569 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:52Z","lastTransitionTime":"2026-01-23T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.092024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.092063 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.092072 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.092085 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.092094 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.194684 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.194750 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.194767 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.194790 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.194808 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.298464 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.298509 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.298537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.298552 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.298560 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.400242 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.400294 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.400310 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.400331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.400344 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.503072 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.503109 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.503119 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.503135 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.503147 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.510942 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 21:47:58.263904097 +0000 UTC Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.514348 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.514374 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.514403 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.514369 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:53 crc kubenswrapper[4899]: E0123 08:52:53.514489 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:53 crc kubenswrapper[4899]: E0123 08:52:53.514580 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:53 crc kubenswrapper[4899]: E0123 08:52:53.514779 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:53 crc kubenswrapper[4899]: E0123 08:52:53.514878 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.533986 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.551382 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.565790 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.581809 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.594602 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606380 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606615 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606629 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.606663 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.618821 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.631755 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.644409 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.654160 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.671039 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.682438 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.692658 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.702358 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.709374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.709409 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.709421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.709436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.709446 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.713336 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.734646 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.745807 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:53Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.811973 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.812039 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.812056 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.812079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.812096 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.914587 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.914635 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.914646 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.914661 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:53 crc kubenswrapper[4899]: I0123 08:52:53.914672 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:53Z","lastTransitionTime":"2026-01-23T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.018665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.018713 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.018723 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.018738 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.018749 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.121100 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.121399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.121410 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.121430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.121442 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.224262 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.224320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.224336 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.224357 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.224373 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.327280 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.327329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.327340 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.327356 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.327367 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.430387 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.430432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.430449 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.430470 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.430525 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.511350 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 20:36:54.36125205 +0000 UTC Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.533382 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.533409 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.533418 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.533430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.533439 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.636023 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.636077 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.636095 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.636120 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.636136 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.738783 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.738824 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.738832 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.738845 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.738856 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.841374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.841408 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.841417 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.841430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.841438 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.944368 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.944419 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.944431 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.944447 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:54 crc kubenswrapper[4899]: I0123 08:52:54.944459 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:54Z","lastTransitionTime":"2026-01-23T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.047173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.047309 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.047337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.047373 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.047398 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.150292 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.150337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.150346 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.150369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.150386 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.253386 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.253438 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.253448 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.253466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.253480 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.356533 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.356606 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.356639 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.356874 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.356901 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.459150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.459202 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.459251 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.459284 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.459305 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.512493 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 03:58:40.218317647 +0000 UTC Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.513895 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.513956 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.513958 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:55 crc kubenswrapper[4899]: E0123 08:52:55.514098 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.514131 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:55 crc kubenswrapper[4899]: E0123 08:52:55.514291 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:55 crc kubenswrapper[4899]: E0123 08:52:55.514308 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:55 crc kubenswrapper[4899]: E0123 08:52:55.514375 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.561836 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.561901 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.561920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.561943 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.561961 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.664085 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.664138 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.664154 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.664538 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.664583 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.767171 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.767274 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.767298 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.767325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.767367 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.869962 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.870022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.870031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.870068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.870078 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.972744 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.972773 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.972781 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.972793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:55 crc kubenswrapper[4899]: I0123 08:52:55.972801 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:55Z","lastTransitionTime":"2026-01-23T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.074627 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.074650 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.074658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.074671 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.074679 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.177106 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.177311 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.177325 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.177376 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.177407 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.280788 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.280836 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.280848 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.280864 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.280875 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.383014 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.383057 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.383068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.383084 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.383094 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.485150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.485236 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.485246 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.485260 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.485288 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.512816 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 10:46:10.860788329 +0000 UTC Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.587743 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.587793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.587804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.587827 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.587850 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.691934 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.691980 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.692031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.692049 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.692065 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.794774 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.794820 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.794831 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.794849 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.794860 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.897428 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.897463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.897473 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.897487 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:56 crc kubenswrapper[4899]: I0123 08:52:56.897497 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:56Z","lastTransitionTime":"2026-01-23T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.000056 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.000101 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.000111 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.000128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.000139 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.102583 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.102617 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.102625 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.102638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.102648 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.205310 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.205388 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.205400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.205417 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.205429 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.307550 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.307585 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.307595 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.307610 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.307618 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.410537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.410601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.410619 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.410645 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.410658 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.512961 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 01:18:51.855028238 +0000 UTC Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513345 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513353 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513366 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513374 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513554 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513556 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513610 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.513664 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:57 crc kubenswrapper[4899]: E0123 08:52:57.513684 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:57 crc kubenswrapper[4899]: E0123 08:52:57.513754 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:57 crc kubenswrapper[4899]: E0123 08:52:57.513837 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:57 crc kubenswrapper[4899]: E0123 08:52:57.513913 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.615910 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.615957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.615973 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.615994 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.616008 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.717793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.717825 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.717833 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.717847 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.717856 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.820252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.820294 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.820303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.820320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.820334 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.922748 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.922791 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.922799 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.922813 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.922821 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.982647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.982699 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.982709 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.982721 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:57 crc kubenswrapper[4899]: I0123 08:52:57.982731 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:57Z","lastTransitionTime":"2026-01-23T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:57 crc kubenswrapper[4899]: E0123 08:52:57.998564 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:57Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.002144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.002198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.002241 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.002256 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.002266 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: E0123 08:52:58.014654 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:58Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.017909 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.018018 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.018085 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.018149 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.018221 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: E0123 08:52:58.029271 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:58Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.032396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.032450 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.032478 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.032492 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.032501 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: E0123 08:52:58.043043 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:58Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.046286 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.046317 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.046328 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.046346 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.046357 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: E0123 08:52:58.058577 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:58Z is after 2025-08-24T17:21:41Z" Jan 23 08:52:58 crc kubenswrapper[4899]: E0123 08:52:58.058742 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.060951 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.061088 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.061172 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.061335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.061426 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.163028 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.163071 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.163082 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.163100 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.163112 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.265945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.265983 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.265994 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.266010 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.266022 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.368173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.368429 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.368506 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.368583 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.368648 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.471429 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.471682 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.471782 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.471866 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.471964 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.513551 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 22:45:42.203497522 +0000 UTC Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.574656 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.574706 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.574723 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.574739 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.574751 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.676807 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.676859 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.676870 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.676889 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.676901 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.778819 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.778865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.778879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.778900 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.778915 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.881132 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.881446 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.881516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.881580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.881640 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.984186 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.984502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.984580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.984643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:58 crc kubenswrapper[4899]: I0123 08:52:58.984711 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:58Z","lastTransitionTime":"2026-01-23T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.087642 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.087694 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.087702 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.087717 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.087726 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.190130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.190176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.190188 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.190218 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.190227 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.292716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.292756 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.292768 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.292784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.292797 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.394395 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.394659 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.394762 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.394895 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.394992 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.497061 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.497106 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.497117 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.497133 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.497146 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.513511 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.513552 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.513537 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.513516 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:52:59 crc kubenswrapper[4899]: E0123 08:52:59.513666 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.513674 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 06:28:42.500162486 +0000 UTC Jan 23 08:52:59 crc kubenswrapper[4899]: E0123 08:52:59.513725 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:52:59 crc kubenswrapper[4899]: E0123 08:52:59.513857 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:52:59 crc kubenswrapper[4899]: E0123 08:52:59.513934 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.599625 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.599679 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.599696 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.599711 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.599719 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.701675 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.701714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.701726 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.701741 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.701753 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.804718 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.804760 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.804771 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.804788 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.804801 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.906957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.907011 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.907022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.907040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:52:59 crc kubenswrapper[4899]: I0123 08:52:59.907050 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:52:59Z","lastTransitionTime":"2026-01-23T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.009962 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.010004 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.010017 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.010034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.010045 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.112112 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.112152 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.112160 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.112176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.112184 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.214316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.214364 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.214380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.214404 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.214419 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.317680 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.317717 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.317729 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.317744 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.317754 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.420063 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.420130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.420140 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.420155 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.420187 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.514450 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 21:39:14.84017088 +0000 UTC Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.522040 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.522078 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.522088 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.522102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.522111 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.624421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.624466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.624479 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.624496 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.624507 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.727031 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.727064 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.727073 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.727088 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.727102 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.829753 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.829824 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.829833 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.829848 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.829856 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.931862 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.931905 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.931916 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.931933 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:00 crc kubenswrapper[4899]: I0123 08:53:00.931946 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:00Z","lastTransitionTime":"2026-01-23T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.034182 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.034238 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.034249 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.034263 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.034272 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.137407 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.137442 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.137456 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.137472 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.137482 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.239056 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.239086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.239096 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.239109 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.239123 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.341025 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.341058 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.341068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.341082 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.341092 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.443639 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.443688 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.443699 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.443715 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.443726 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.513964 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.513997 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.513980 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.514314 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:01 crc kubenswrapper[4899]: E0123 08:53:01.514445 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:01 crc kubenswrapper[4899]: E0123 08:53:01.514471 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:01 crc kubenswrapper[4899]: E0123 08:53:01.514537 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:01 crc kubenswrapper[4899]: E0123 08:53:01.514583 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.514607 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 16:17:54.438191253 +0000 UTC Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.514777 4899 scope.go:117] "RemoveContainer" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" Jan 23 08:53:01 crc kubenswrapper[4899]: E0123 08:53:01.515149 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.546039 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.546084 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.546096 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.546115 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.546127 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.648913 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.648955 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.648965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.648984 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.648996 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.751832 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.752033 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.752091 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.752183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.752260 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.854310 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.854348 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.854359 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.854374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.854390 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.956746 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.956792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.956811 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.956832 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:01 crc kubenswrapper[4899]: I0123 08:53:01.956845 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:01Z","lastTransitionTime":"2026-01-23T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.059103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.059149 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.059160 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.059178 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.059190 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.161084 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.161116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.161128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.161144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.161155 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.263047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.263081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.263093 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.263108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.263118 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.365405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.365439 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.365448 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.365462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.365471 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.467398 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.467458 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.467474 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.467493 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.467510 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.514704 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:25:13.664859742 +0000 UTC Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.570169 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.570255 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.570294 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.570329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.570351 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.673663 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.673709 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.673719 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.673735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.673746 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.776394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.776450 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.776466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.776488 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.776504 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.880018 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.880070 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.880082 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.880100 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.880112 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.983137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.983188 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.983227 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.983247 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:02 crc kubenswrapper[4899]: I0123 08:53:02.983270 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:02Z","lastTransitionTime":"2026-01-23T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.086182 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.086261 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.086297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.086368 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.086381 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.188298 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.188326 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.188334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.188346 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.188355 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.290751 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.290785 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.290794 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.290809 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.290818 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.393024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.393081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.393097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.393122 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.393139 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.495192 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.495268 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.495284 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.495303 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.495316 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.514702 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.514751 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.514808 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 22:26:14.451331232 +0000 UTC Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.514728 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.514855 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.514985 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.515019 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.515053 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.515104 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.534743 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.556729 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.570557 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.584452 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.596298 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.597252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.597375 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.597626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.597710 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.597789 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.611597 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.635993 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.650462 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.663282 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.675332 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.686437 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.696305 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.699483 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.699516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.699526 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.699542 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.699554 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.706648 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.717322 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.728781 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.739351 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.748265 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:03Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.805078 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.805396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.805490 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.805589 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.805666 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.907616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.907657 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.907670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.907684 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.907694 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:03Z","lastTransitionTime":"2026-01-23T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:03 crc kubenswrapper[4899]: I0123 08:53:03.976460 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.976605 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:53:03 crc kubenswrapper[4899]: E0123 08:53:03.976659 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:53:35.976645221 +0000 UTC m=+102.843615608 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.010156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.010189 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.010197 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.010223 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.010232 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.112382 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.112584 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.112683 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.112750 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.112810 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.214430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.214692 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.214756 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.214816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.214877 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.316702 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.316746 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.316755 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.316770 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.316779 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.419807 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.419865 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.419878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.419897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.419912 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.515089 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 16:38:49.48857733 +0000 UTC Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.522544 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.522593 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.522606 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.522626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.522636 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.625265 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.625349 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.625364 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.625390 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.625406 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.728134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.728174 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.728182 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.728196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.728225 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.830719 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.830761 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.830773 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.830792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.830804 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.928634 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/0.log" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.928680 4899 generic.go:334] "Generic (PLEG): container finished" podID="60a84ebf-a188-4eee-8a9f-e02ac7adb8c8" containerID="e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6" exitCode=1 Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.928708 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerDied","Data":"e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.929032 4899 scope.go:117] "RemoveContainer" containerID="e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.933087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.933108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.933116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.933130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.933139 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:04Z","lastTransitionTime":"2026-01-23T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.945138 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:04Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.960073 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:04Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.974365 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:04Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.987120 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:04Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:04 crc kubenswrapper[4899]: I0123 08:53:04.998910 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:04Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.013608 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.026540 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.035610 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.035660 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.035673 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.035702 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.035716 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.039385 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.052854 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.069130 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.080880 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.093939 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.106940 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.123314 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.134933 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.137583 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.137613 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.137638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.137657 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.137669 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.158174 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.170706 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.239906 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.240141 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.240232 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.240338 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.240415 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.343176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.343350 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.343421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.343485 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.343541 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.447065 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.447116 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.447130 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.447152 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.447167 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.514184 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.514262 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.514319 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.514346 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:05 crc kubenswrapper[4899]: E0123 08:53:05.514977 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.515475 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 18:08:01.471940258 +0000 UTC Jan 23 08:53:05 crc kubenswrapper[4899]: E0123 08:53:05.515726 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:05 crc kubenswrapper[4899]: E0123 08:53:05.516099 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:05 crc kubenswrapper[4899]: E0123 08:53:05.516272 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.549601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.549655 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.549667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.549686 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.549702 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.652368 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.652427 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.652443 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.652466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.652482 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.755150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.755405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.755534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.755638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.755736 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.858319 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.858363 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.858374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.858392 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.858404 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.933492 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/0.log" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.933798 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerStarted","Data":"cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.947574 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.958510 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.960696 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.960747 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.960759 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.960779 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.960790 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:05Z","lastTransitionTime":"2026-01-23T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.969732 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.979553 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:05 crc kubenswrapper[4899]: I0123 08:53:05.990755 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:05Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.003492 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.018973 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.030257 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.044264 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.054365 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.063464 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.063670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.063793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.063932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.064049 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.065429 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.080916 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.090014 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.100495 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.111110 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.124127 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.138149 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:06Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.166803 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.166841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.166879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.166897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.166908 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.268711 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.268751 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.268762 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.268779 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.268789 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.371279 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.371557 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.371643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.371734 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.371809 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.474250 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.474286 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.474296 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.474313 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.474324 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.515724 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 23:02:09.190061553 +0000 UTC Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.577033 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.577073 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.577081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.577095 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.577106 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.679289 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.679350 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.679373 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.679400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.679423 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.782005 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.782047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.782059 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.782074 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.782084 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.884866 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.885203 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.885364 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.885469 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.885553 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.988334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.988381 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.988392 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.988408 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:06 crc kubenswrapper[4899]: I0123 08:53:06.988419 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:06Z","lastTransitionTime":"2026-01-23T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.090776 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.090826 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.090837 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.091024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.091037 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.194720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.195071 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.195424 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.195518 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.195592 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.299034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.299111 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.299133 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.299162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.299186 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.401883 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.401931 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.401942 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.401958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.401968 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.504667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.504725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.504740 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.504761 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.504776 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.514088 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.514175 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:07 crc kubenswrapper[4899]: E0123 08:53:07.514239 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.514292 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:07 crc kubenswrapper[4899]: E0123 08:53:07.514428 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.514615 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:07 crc kubenswrapper[4899]: E0123 08:53:07.514722 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:07 crc kubenswrapper[4899]: E0123 08:53:07.514784 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.516139 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 18:15:20.40322156 +0000 UTC Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.606548 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.606582 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.606592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.606607 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.606617 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.708706 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.708935 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.709011 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.709074 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.709141 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.812003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.812402 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.812511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.812644 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.812740 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.915728 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.915772 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.915782 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.915794 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:07 crc kubenswrapper[4899]: I0123 08:53:07.915802 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:07Z","lastTransitionTime":"2026-01-23T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.018977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.019011 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.019022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.019038 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.019049 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.108878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.108921 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.108932 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.108948 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.108959 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.119792 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:08Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.123451 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.123502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.123512 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.123525 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.123534 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.134317 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:08Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.136911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.136940 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.136948 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.136960 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.136969 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.150304 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:08Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.154566 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.154604 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.154616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.154658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.154677 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.166733 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:08Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.169087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.169112 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.169123 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.169136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.169147 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.180015 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:08Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:08 crc kubenswrapper[4899]: E0123 08:53:08.180166 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.181556 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.181582 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.181592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.181604 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.181612 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.285082 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.285113 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.285121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.285134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.285149 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.388327 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.388410 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.388436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.388473 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.388491 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.492358 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.492395 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.492405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.492420 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.492429 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.517129 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 16:04:26.253520172 +0000 UTC Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.594666 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.594701 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.594709 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.594724 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.594733 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.696396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.696463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.696507 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.696537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.696558 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.799253 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.799320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.799331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.799347 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.799357 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.901097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.901151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.901166 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.901182 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:08 crc kubenswrapper[4899]: I0123 08:53:08.901194 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:08Z","lastTransitionTime":"2026-01-23T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.009499 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.009553 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.009568 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.009587 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.009620 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.113079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.113131 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.113142 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.113158 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.113168 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.216626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.216700 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.216724 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.216751 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.216773 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.320816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.320887 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.320925 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.320955 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.320974 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.424247 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.424291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.424300 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.424319 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.424332 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.514118 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.514152 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.514471 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.514576 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:09 crc kubenswrapper[4899]: E0123 08:53:09.514594 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:09 crc kubenswrapper[4899]: E0123 08:53:09.514724 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:09 crc kubenswrapper[4899]: E0123 08:53:09.514806 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:09 crc kubenswrapper[4899]: E0123 08:53:09.514872 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.517698 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 23:30:33.375394955 +0000 UTC Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.527047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.527129 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.527150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.527181 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.527252 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.630919 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.630993 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.631014 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.631043 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.631067 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.733920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.733963 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.733976 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.733994 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.734005 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.837798 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.837842 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.837856 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.837871 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.837883 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.940300 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.940379 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.940400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.940430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:09 crc kubenswrapper[4899]: I0123 08:53:09.940451 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:09Z","lastTransitionTime":"2026-01-23T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.043608 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.043658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.043675 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.043701 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.043718 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.146086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.146127 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.146136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.146150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.146161 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.248922 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.248981 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.248992 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.249008 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.249019 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.351607 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.351659 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.351674 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.351695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.351707 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.454335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.454374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.454381 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.454412 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.454423 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.518238 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 03:55:30.867616693 +0000 UTC Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.526721 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.556248 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.556302 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.556314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.556330 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.556341 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.659588 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.659638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.659650 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.659667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.659678 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.761845 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.761908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.761924 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.761943 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.761957 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.864027 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.864139 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.864162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.864190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.864242 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.967450 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.967516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.967532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.967558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:10 crc kubenswrapper[4899]: I0123 08:53:10.967573 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:10Z","lastTransitionTime":"2026-01-23T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.070745 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.070832 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.070851 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.070878 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.070896 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.173765 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.173822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.173841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.173862 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.173877 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.276177 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.276239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.276251 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.276267 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.276277 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.378436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.378471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.378483 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.378498 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.378509 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.482778 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.482876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.482900 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.482928 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.482945 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.514596 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.514686 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:11 crc kubenswrapper[4899]: E0123 08:53:11.514750 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.514714 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:11 crc kubenswrapper[4899]: E0123 08:53:11.514979 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.515002 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:11 crc kubenswrapper[4899]: E0123 08:53:11.515082 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:11 crc kubenswrapper[4899]: E0123 08:53:11.515153 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.518646 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:05:37.653982801 +0000 UTC Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.585632 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.585693 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.585710 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.586078 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.586137 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.690413 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.690491 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.690518 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.690547 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.690570 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.793517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.793593 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.793617 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.793647 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.793668 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.896910 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.896997 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.897022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.897055 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.897089 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.999672 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.999716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.999741 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.999764 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:11 crc kubenswrapper[4899]: I0123 08:53:11.999778 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:11Z","lastTransitionTime":"2026-01-23T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.103146 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.103248 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.103273 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.103299 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.103319 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.206055 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.206133 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.206156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.206184 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.206258 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.309545 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.309590 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.309599 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.309617 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.309629 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.411683 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.411727 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.411739 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.411755 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.411766 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514292 4899 scope.go:117] "RemoveContainer" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514412 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514423 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514437 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.514449 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.519457 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:25:08.330437254 +0000 UTC Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.616994 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.617026 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.617034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.617046 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.617054 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.721554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.721665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.721687 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.721714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.721733 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.824425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.824476 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.824490 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.824515 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.824531 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.927580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.927621 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.927633 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.927650 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.927663 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:12Z","lastTransitionTime":"2026-01-23T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.957364 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/2.log" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.959452 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.961020 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.975093 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:12Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:12 crc kubenswrapper[4899]: I0123 08:53:12.998593 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:12Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.010293 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.030157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.030196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.030225 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.030243 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.030255 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.032333 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.051638 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.062841 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"389f49af-b0cf-49f2-885d-5ee190be236b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.074324 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.093821 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.104887 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.117873 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.127284 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.132536 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.132565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.132574 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.132587 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.132597 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.137692 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.150272 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.162580 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.176517 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.187372 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.199817 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.211692 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.234674 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.234720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.234728 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.234741 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.234750 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.336822 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.336864 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.336874 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.336889 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.336900 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.439507 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.439551 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.439560 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.439575 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.439584 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.514297 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:13 crc kubenswrapper[4899]: E0123 08:53:13.514407 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.514459 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.514560 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:13 crc kubenswrapper[4899]: E0123 08:53:13.514587 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.514645 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:13 crc kubenswrapper[4899]: E0123 08:53:13.514729 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:13 crc kubenswrapper[4899]: E0123 08:53:13.514779 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.521331 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 11:11:58.748929382 +0000 UTC Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.530306 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.539361 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.541635 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.541665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.541674 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.541686 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.541695 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.549923 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.564185 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.576662 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.585455 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.599576 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.611400 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.621222 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"389f49af-b0cf-49f2-885d-5ee190be236b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.632549 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643533 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643720 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643811 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643838 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643864 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.643883 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.655058 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.674514 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.685895 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.698853 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.713739 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.727972 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.738579 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.746222 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.746260 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.746273 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.746295 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.746309 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.849225 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.849271 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.849282 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.849297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.849312 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.952758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.952806 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.952816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.952830 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.952840 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:13Z","lastTransitionTime":"2026-01-23T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.963871 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/3.log" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.964434 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/2.log" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.967115 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" exitCode=1 Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.967155 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.967187 4899 scope.go:117] "RemoveContainer" containerID="3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.968454 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:53:13 crc kubenswrapper[4899]: E0123 08:53:13.970300 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.981520 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:13 crc kubenswrapper[4899]: I0123 08:53:13.993816 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:13Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.007019 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.017130 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.030392 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.042928 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.055331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.055384 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.055396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.055415 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.055425 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.056274 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"389f49af-b0cf-49f2-885d-5ee190be236b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.066386 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.083075 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3864c8437c35829bd7a990af60d84cd9be554ba1ca584f39df9ff5f9044cded7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:52:41Z\\\",\\\"message\\\":\\\"09437 6597 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:52:41Z is after 2025-08-24T17:21:41Z]\\\\nI0123 08:52:41.409393 6597 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:13Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:53:13.477324 7016 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:53:13.477358 7016 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:53:13.477363 7016 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:53:13.477383 7016 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:53:13.477418 7016 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:53:13.477428 7016 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0123 08:53:13.477433 7016 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0123 08:53:13.477456 7016 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:53:13.477471 7016 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:53:13.477468 7016 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:53:13.477477 7016 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:53:13.477489 7016 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:53:13.477493 7016 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:53:13.477511 7016 factory.go:656] Stopping watch factory\\\\nI0123 08:53:13.477524 7016 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.095803 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.108576 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.120110 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.133493 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.144805 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.156728 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.157411 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.157443 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.157460 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.157479 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.157491 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.170375 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.184343 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.194542 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.260010 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.260053 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.260065 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.260080 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.260090 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.363198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.363252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.363262 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.363276 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.363288 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.465673 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.465703 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.465713 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.465728 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.465738 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.521873 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 01:36:20.348849288 +0000 UTC Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.567920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.567958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.567983 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.568003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.568017 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.670149 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.670406 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.670444 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.670473 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.670493 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.773389 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.773444 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.773456 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.773470 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.773481 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.875258 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.875292 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.875300 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.875315 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.875324 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.972233 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/3.log" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.975135 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:53:14 crc kubenswrapper[4899]: E0123 08:53:14.975362 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.977355 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.977378 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.977386 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.977399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.977409 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:14Z","lastTransitionTime":"2026-01-23T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:14 crc kubenswrapper[4899]: I0123 08:53:14.988320 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:14Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.002048 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.015106 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.028531 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.040594 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.054541 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.064804 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.075149 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.078982 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.079015 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.079023 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.079036 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.079047 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.085737 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.096457 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.110003 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.120488 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.136114 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.147854 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.157548 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"389f49af-b0cf-49f2-885d-5ee190be236b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.172871 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.180903 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.180935 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.180943 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.180957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.180965 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.189425 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:13Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:53:13.477324 7016 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:53:13.477358 7016 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:53:13.477363 7016 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:53:13.477383 7016 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:53:13.477418 7016 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:53:13.477428 7016 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0123 08:53:13.477433 7016 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0123 08:53:13.477456 7016 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:53:13.477471 7016 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:53:13.477468 7016 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:53:13.477477 7016 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:53:13.477489 7016 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:53:13.477493 7016 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:53:13.477511 7016 factory.go:656] Stopping watch factory\\\\nI0123 08:53:13.477524 7016 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.197611 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:15Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.284023 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.284060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.284070 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.284087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.284098 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.386349 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.386412 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.386427 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.386452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.386468 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.488804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.488874 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.488891 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.488918 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.488939 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.514047 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.514109 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.514169 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.514231 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:15 crc kubenswrapper[4899]: E0123 08:53:15.514238 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:15 crc kubenswrapper[4899]: E0123 08:53:15.514402 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:15 crc kubenswrapper[4899]: E0123 08:53:15.515004 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:15 crc kubenswrapper[4899]: E0123 08:53:15.515167 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.522765 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 08:23:06.827014211 +0000 UTC Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.535584 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.592916 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.592996 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.593015 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.593041 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.593060 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.696414 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.696891 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.697050 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.697248 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.697410 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.800250 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.800880 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.801119 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.801369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.801573 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.905769 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.905837 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.905859 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.905885 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:15 crc kubenswrapper[4899]: I0123 08:53:15.905904 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:15Z","lastTransitionTime":"2026-01-23T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.008968 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.009024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.009037 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.009085 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.009099 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.111784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.111861 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.111885 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.111914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.111937 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.215350 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.215394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.215415 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.215474 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.215493 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.317930 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.317993 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.318009 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.318033 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.318142 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.420394 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.420455 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.420470 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.420494 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.420509 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.522874 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 23:39:53.531171013 +0000 UTC Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.523491 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.523568 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.523592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.523622 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.523646 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.626191 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.626276 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.626285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.626298 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.626309 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.729519 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.729596 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.729621 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.729652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.729674 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.832354 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.832393 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.832402 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.832417 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.832427 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.935060 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.935125 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.935147 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.935174 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:16 crc kubenswrapper[4899]: I0123 08:53:16.935194 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:16Z","lastTransitionTime":"2026-01-23T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.037366 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.037432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.037457 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.037482 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.037499 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.140290 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.140379 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.140396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.140421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.140437 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.248805 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.248879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.248897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.248924 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.248942 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.332937 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.333245 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.333188495 +0000 UTC m=+148.200158892 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.351867 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.351900 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.351911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.351927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.351938 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.433970 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.434055 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.434108 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.434281 4899 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.434388 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.434356874 +0000 UTC m=+148.301327291 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.434535 4899 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.434645 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.434618501 +0000 UTC m=+148.301588928 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.434800 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435094 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435178 4899 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435324 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.435296488 +0000 UTC m=+148.302266915 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.435430 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435555 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435594 4899 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435616 4899 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.435729 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.435706858 +0000 UTC m=+148.302677335 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.453911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.453943 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.453953 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.453965 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.453974 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.513820 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.513845 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.513927 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.513955 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.514068 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.514300 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.514422 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:17 crc kubenswrapper[4899]: E0123 08:53:17.514633 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.523013 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 02:56:15.997047496 +0000 UTC Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.556268 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.556323 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.556339 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.556362 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.556380 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.659477 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.659527 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.659544 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.659567 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.659584 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.765700 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.765769 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.765784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.765810 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.765826 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.868592 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.868649 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.868661 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.868678 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.868689 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.971549 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.971610 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.971629 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.971651 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:17 crc kubenswrapper[4899]: I0123 08:53:17.971669 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:17Z","lastTransitionTime":"2026-01-23T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.074731 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.074840 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.074860 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.074884 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.074902 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.178109 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.178179 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.178202 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.178289 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.178322 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.280873 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.281340 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.281441 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.281589 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.281675 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.384555 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.384620 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.384643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.384673 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.384695 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.487585 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.487705 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.487729 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.487756 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.487774 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.523658 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 11:06:16.468254905 +0000 UTC Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.573112 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.573336 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.573351 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.573369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.573383 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.585572 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.588517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.588556 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.588569 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.588586 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.588596 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.599140 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.603396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.603424 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.603432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.603447 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.603456 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.613708 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.616450 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.616493 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.616503 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.616519 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.616533 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.629449 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.632612 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.632667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.632681 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.632695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.632704 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.643938 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:18Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:18 crc kubenswrapper[4899]: E0123 08:53:18.644094 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.645771 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.645807 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.645815 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.645831 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.645841 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.752361 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.752392 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.752400 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.752414 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.752423 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.855042 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.855432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.855584 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.855735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.855881 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.959049 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.959090 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.959100 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.959117 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:18 crc kubenswrapper[4899]: I0123 08:53:18.959128 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:18Z","lastTransitionTime":"2026-01-23T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.062584 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.063062 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.063286 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.063509 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.064166 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.168258 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.168652 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.168823 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.168993 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.169148 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.272597 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.273155 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.273696 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.274134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.275040 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.378480 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.378537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.378549 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.378566 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.378579 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.481445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.481490 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.481504 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.481525 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.481538 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.514137 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.514152 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.514161 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:19 crc kubenswrapper[4899]: E0123 08:53:19.514621 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:19 crc kubenswrapper[4899]: E0123 08:53:19.514479 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.514310 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:19 crc kubenswrapper[4899]: E0123 08:53:19.514906 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:19 crc kubenswrapper[4899]: E0123 08:53:19.514936 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.524929 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 14:52:40.350118631 +0000 UTC Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.584382 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.584430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.584443 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.584462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.584476 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.688047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.688430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.688576 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.688709 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.688802 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.792057 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.792102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.792111 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.792128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.792140 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.894355 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.894422 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.894440 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.894465 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.894482 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.997343 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.997611 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.997735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.997863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:19 crc kubenswrapper[4899]: I0123 08:53:19.997987 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:19Z","lastTransitionTime":"2026-01-23T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.100601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.100658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.100672 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.100703 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.100717 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.203684 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.204047 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.204153 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.204277 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.204371 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.306938 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.307001 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.307012 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.307024 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.307033 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.415041 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.415074 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.415083 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.415097 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.415111 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.518556 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.518610 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.518626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.518650 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.518668 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.526026 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 20:07:41.460284828 +0000 UTC Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.621725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.621796 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.621812 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.621841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.621856 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.725532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.725606 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.725617 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.725640 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.725654 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.829268 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.829332 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.829343 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.829366 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.829384 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.932793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.932880 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.932893 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.932918 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:20 crc kubenswrapper[4899]: I0123 08:53:20.932940 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:20Z","lastTransitionTime":"2026-01-23T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.036541 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.036598 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.036616 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.036640 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.036658 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.139612 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.139682 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.139703 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.139730 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.139748 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.242230 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.242275 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.242285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.242302 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.242313 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.345072 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.345182 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.345253 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.345275 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.345293 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.448277 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.448326 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.448338 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.448356 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.448369 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.513596 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.513685 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.513762 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.513793 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:21 crc kubenswrapper[4899]: E0123 08:53:21.514142 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:21 crc kubenswrapper[4899]: E0123 08:53:21.514313 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:21 crc kubenswrapper[4899]: E0123 08:53:21.514001 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:21 crc kubenswrapper[4899]: E0123 08:53:21.514618 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.526853 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 07:13:05.371926071 +0000 UTC Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.551034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.551065 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.551081 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.551099 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.551110 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.653297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.653372 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.653424 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.653441 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.653454 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.756500 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.756568 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.756586 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.756601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.756609 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.859731 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.859779 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.859795 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.859811 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.859824 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.964265 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.964367 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.964396 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.964429 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:21 crc kubenswrapper[4899]: I0123 08:53:21.964447 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:21Z","lastTransitionTime":"2026-01-23T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.067543 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.067591 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.067602 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.067624 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.067636 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.170344 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.170418 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.170436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.170462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.170480 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.274066 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.274149 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.274175 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.274205 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.274300 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.377452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.377512 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.377535 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.377568 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.377594 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.481507 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.481604 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.481643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.481669 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.481686 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.527655 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 15:53:15.69273672 +0000 UTC Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.584370 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.584433 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.584458 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.584488 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.584508 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.687785 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.687862 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.687887 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.687953 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.687982 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.790079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.790137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.790158 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.790179 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.790195 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.892909 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.892949 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.892958 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.892975 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.892984 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.996073 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.996114 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.996128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.996145 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:22 crc kubenswrapper[4899]: I0123 08:53:22.996159 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:22Z","lastTransitionTime":"2026-01-23T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.098758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.098835 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.098855 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.098879 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.098896 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.202983 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.203042 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.203062 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.203086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.203105 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.305735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.305790 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.305806 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.305826 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.305840 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.408108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.408156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.408173 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.408194 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.408228 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.511510 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.511574 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.511593 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.511646 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.511668 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.513863 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.513901 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:23 crc kubenswrapper[4899]: E0123 08:53:23.514032 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.514122 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.514259 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:23 crc kubenswrapper[4899]: E0123 08:53:23.514476 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:23 crc kubenswrapper[4899]: E0123 08:53:23.514562 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:23 crc kubenswrapper[4899]: E0123 08:53:23.514658 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.528547 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 06:51:47.47911781 +0000 UTC Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.550665 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b81bdf1a-b8e5-47cf-8141-839449b59a2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de86be5abee9338d5508816dc4f77e379f1b14a70a4254e79174c608e282a242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54e65a53617c7eeeb2bbd234503ed2546384859d8fcc727f11e26db2744b100f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4c1a320817528fef7347af28c56b752dd79cd884c15441aa33bfedea374b48c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd057062f7680519fe8fee23eabb0d9994b79474b9c46e3293cd79220d95684e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b26664c1dbf488573e455beba962306bf6750a2d359e4b3d222b24f7816331cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23dbb59d872d663a876e9d320bd90e3ee539f107377299991998b285a63b10b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23dbb59d872d663a876e9d320bd90e3ee539f107377299991998b285a63b10b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f736c2fc78cb296e65f67467c6cce91ce967652ae2b67eebf638d999ff3178eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f736c2fc78cb296e65f67467c6cce91ce967652ae2b67eebf638d999ff3178eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://83bc211613c8889b1c214d3503fb83ee7b0ab2f0543fcd95b784ebe4b601d86b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83bc211613c8889b1c214d3503fb83ee7b0ab2f0543fcd95b784ebe4b601d86b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.569140 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"140c9bb6-a73d-47fa-b655-372ff3451f26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b409fc1fffdf9ca51482abeb11d257327191972f1d5e5ebd85821d0ad0c9e87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5816dbfda12a244173271a21c96323d90c907a9a152a8560c198ee765e770f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba85001d97facd17d84c26e63fab6f056dd13eaa6c57dd25354b7c79fdc4954\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd69897d3d01ad36c62479f7e3a95803c0cb456410f40006233d437a104f3fb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.586891 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.607106 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5f4c0354a0d8af6b8b0010312cc384fe3c0967900ded7581a58d68b2c1716aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.614942 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.614988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.615000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.615017 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.615029 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.627999 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.644450 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb844e06-c862-406b-8ecb-36b682d2d58b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7l67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxcxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.658922 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"389f49af-b0cf-49f2-885d-5ee190be236b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfb8e0367bd9b89222d7e094bce0d11ff3562f52f1cb765fab51c870bc613dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e053df2e1dd92cede4f6658e3cad97d7f73282fa00449f92e4402a3ea4355247\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.676741 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.697630 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09d03f16533cd667a7dee4d4cf6eeb050f6df9aec74cd39eb3428cc4b26a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718353 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718424 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718454 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718466 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.718941 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737f0e4e072f6423b70719bad88d58995b0b704f69459a8137b335e66776c804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec3bd5edb15bc7a6e9d5b4306c419487ad950cb872804ecc38655e073f6647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.732417 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nzqf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7629148-8035-4115-b7f7-89c2868cad0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329d687463973fcb8c145a6efff49ee41493e702237050c72fd14a85123de656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sjqs7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nzqf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.753363 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9a9cb2-ec5b-4498-ab3e-0061bc62e7b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71767d5b82418b7c46b39ec7d22bbfa38b5cc6187c5030d0fb2bd932e8223c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5362a1893aa7fc518cdd45989600b4328f5fcc81b26d2869418965258aa391c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76c8d35297e9dcd762557854f19a38c429459e3257ed26fead0b49ce7234ea4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f71fc496384f77cb37887bf091c257f44931a5b43b3df037b678a76d78cdc824\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99bc46fd5d4885d5aea4e7fcc0c15aa1b7cda866870755f1def285135093e25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dab7faea232d8a6b93e3a0010c1a31749bd141f29dbe19c09b8967167229354\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caad76f7db72bf93e6158196fa9396db5f85220f922351c6afad6fec8ee2e36c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxsd6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gbqr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.765873 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28c6ba1e-01d3-4b6f-b539-5279a6cbfed0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac4818e350d6d1db34623048becd3ab93613c1e37cc17c989b2962e56ad4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://504804f408958b084372e1c42f364b268731d5b7a673ccffba9cd831e0dac45b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khwdn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.794866 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"790411d3-0df3-4cc3-bf33-835dab4c2cb7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:13Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0123 08:53:13.477324 7016 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0123 08:53:13.477358 7016 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0123 08:53:13.477363 7016 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0123 08:53:13.477383 7016 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0123 08:53:13.477418 7016 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0123 08:53:13.477428 7016 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0123 08:53:13.477433 7016 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0123 08:53:13.477456 7016 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0123 08:53:13.477471 7016 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0123 08:53:13.477468 7016 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0123 08:53:13.477477 7016 handler.go:208] Removed *v1.Node event handler 2\\\\nI0123 08:53:13.477489 7016 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0123 08:53:13.477493 7016 handler.go:208] Removed *v1.Node event handler 7\\\\nI0123 08:53:13.477511 7016 factory.go:656] Stopping watch factory\\\\nI0123 08:53:13.477524 7016 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4mv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qmqs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.808268 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lmzls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85294630-22ed-44e4-99f6-ae7fa8658dbe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61ac58249ea9c8c5e3811508248048cba6cb3e6d135db99509259308bb571d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5hlq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lmzls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.821120 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.821151 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.821162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.821180 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.821191 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.823128 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c507cbe-e2d0-4289-ab33-df276d77759d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caaf71675f50229b24d20d6fddd601d01f03ba56af5fe358cbe1fe1b605a9d62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a16cbfdf1c3b0d2f11d464c5b5e2aec6676a89451141e51b29fef581de205288\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e922e4c9151951278c6f50fae3a6e2ef9b02e44cc8e752656a9d47d4eb6d88a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.841030 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d29358f2-5909-45b8-bacf-e12b71fea05d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:51:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-23T08:52:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0123 08:52:13.124250 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0123 08:52:13.126110 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0123 08:52:13.127368 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1427814774/tls.crt::/tmp/serving-cert-1427814774/tls.key\\\\\\\"\\\\nI0123 08:52:13.337872 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0123 08:52:13.339590 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0123 08:52:13.339644 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0123 08:52:13.339683 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0123 08:52:13.339708 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0123 08:52:13.343229 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0123 08:52:13.343282 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343288 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0123 08:52:13.343293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0123 08:52:13.343297 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0123 08:52:13.343300 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0123 08:52:13.343304 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0123 08:52:13.343245 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0123 08:52:13.344975 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:51:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-23T08:51:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-23T08:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:51:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.861765 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h5gln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-23T08:53:04Z\\\",\\\"message\\\":\\\"2026-01-23T08:52:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845\\\\n2026-01-23T08:52:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7541f6e2-e2be-4d21-953e-3d114f5c3845 to /host/opt/cni/bin/\\\\n2026-01-23T08:52:19Z [verbose] multus-daemon started\\\\n2026-01-23T08:52:19Z [verbose] Readiness Indicator file check\\\\n2026-01-23T08:53:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sp9g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h5gln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.879453 4899 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57f7d8bc-8a1d-4b44-be6e-d702c34203c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-23T08:52:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbae080c69fae7d53f2e9d4842714cdb40168ee791590823571c338c9eebb866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-23T08:52:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jb8vr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-23T08:52:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6wwr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:23Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.923594 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.923670 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.923688 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.923714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:23 crc kubenswrapper[4899]: I0123 08:53:23.923732 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:23Z","lastTransitionTime":"2026-01-23T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.026342 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.026405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.026423 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.026447 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.026465 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.128636 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.128677 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.128695 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.128714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.128727 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.232057 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.233190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.233512 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.233793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.234058 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.337077 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.337331 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.337405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.337482 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.337552 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.439792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.439845 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.439854 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.439869 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.439881 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.529311 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 02:40:40.069944665 +0000 UTC Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.541933 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.541986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.541997 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.542014 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.542025 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.644715 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.644840 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.644866 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.644894 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.644916 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.746693 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.746754 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.746772 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.746794 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.746814 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.849536 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.849601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.849622 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.849643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.849656 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.952955 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.953064 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.953075 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.953095 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:24 crc kubenswrapper[4899]: I0123 08:53:24.953111 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:24Z","lastTransitionTime":"2026-01-23T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.055380 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.055445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.055459 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.055481 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.055499 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.157945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.158001 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.158016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.158035 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.158052 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.260665 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.260741 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.260754 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.260793 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.260806 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.363335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.363374 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.363382 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.363410 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.363420 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.466575 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.466615 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.466626 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.466641 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.466652 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.514487 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.514562 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.514511 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:25 crc kubenswrapper[4899]: E0123 08:53:25.514694 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.514485 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:25 crc kubenswrapper[4899]: E0123 08:53:25.514920 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:25 crc kubenswrapper[4899]: E0123 08:53:25.514984 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:25 crc kubenswrapper[4899]: E0123 08:53:25.515067 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.530465 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 04:24:36.152648939 +0000 UTC Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.568921 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.568986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.569000 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.569022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.569037 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.671337 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.671373 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.671383 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.671399 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.671411 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.774861 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.774915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.774926 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.774939 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.774951 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.878025 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.878086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.878103 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.878125 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.878142 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.981912 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.982016 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.982045 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.982079 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:25 crc kubenswrapper[4899]: I0123 08:53:25.982101 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:25Z","lastTransitionTime":"2026-01-23T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.085391 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.085462 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.085482 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.085509 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.085527 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.188191 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.188260 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.188273 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.188293 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.188307 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.291310 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.291390 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.291416 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.291445 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.291469 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.394937 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.395067 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.395096 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.395127 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.395154 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.498104 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.498176 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.498191 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.498229 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.498243 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.515684 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:53:26 crc kubenswrapper[4899]: E0123 08:53:26.516017 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.531708 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 17:54:29.064568276 +0000 UTC Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.601754 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.601803 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.601814 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.601832 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.601844 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.705092 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.705136 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.705144 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.705159 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.705171 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.808285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.808336 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.808350 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.808370 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.808383 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.910888 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.910931 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.910959 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.910977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:26 crc kubenswrapper[4899]: I0123 08:53:26.910990 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:26Z","lastTransitionTime":"2026-01-23T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.012921 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.013003 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.013017 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.013048 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.013063 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.115924 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.115993 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.116009 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.116035 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.116056 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.220413 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.220494 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.220511 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.220528 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.220540 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.322951 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.322986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.322994 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.323006 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.323017 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.425481 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.425529 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.425536 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.425552 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.425561 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.513860 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.513960 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.513880 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:27 crc kubenswrapper[4899]: E0123 08:53:27.514050 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.513883 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:27 crc kubenswrapper[4899]: E0123 08:53:27.514178 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:27 crc kubenswrapper[4899]: E0123 08:53:27.514326 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:27 crc kubenswrapper[4899]: E0123 08:53:27.514450 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.527177 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.527247 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.527271 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.527292 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.527343 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.532724 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 03:38:26.140068563 +0000 UTC Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.629507 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.629553 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.629564 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.629581 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.629591 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.732643 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.732711 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.732725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.732741 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.732752 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.835819 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.835867 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.835876 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.835890 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.835899 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.939583 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.942134 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.942154 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.942170 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:27 crc kubenswrapper[4899]: I0123 08:53:27.942181 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:27Z","lastTransitionTime":"2026-01-23T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.044886 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.045086 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.045099 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.045117 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.045129 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.147384 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.147437 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.147451 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.147468 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.147480 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.249873 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.250001 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.250019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.250036 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.250045 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.352391 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.352449 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.352461 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.352482 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.352504 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.455579 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.455636 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.455649 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.455667 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.455681 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.533836 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:25:04.047457898 +0000 UTC Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.558544 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.558574 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.558586 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.558601 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.558611 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.661184 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.661246 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.661260 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.661274 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.661287 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.681378 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.681410 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.681419 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.681433 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.681442 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.692928 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.697735 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.697795 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.697804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.697818 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.697826 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.711420 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.716034 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.716087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.716108 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.716128 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.716139 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.729171 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.732869 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.732897 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.732907 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.732919 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.732927 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.743471 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.746679 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.746708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.746716 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.746728 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.746738 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.757779 4899 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-23T08:53:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"173d46ba-8cf7-44c7-ba1f-ff4b0070b3e5\\\",\\\"systemUUID\\\":\\\"7c35b762-77f6-4705-882d-2520cd67cab9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-23T08:53:28Z is after 2025-08-24T17:21:41Z" Jan 23 08:53:28 crc kubenswrapper[4899]: E0123 08:53:28.757971 4899 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.763685 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.763714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.763725 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.763740 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.763752 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.865702 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.866021 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.866036 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.866056 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.866070 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.968824 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.968875 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.968891 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.968911 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:28 crc kubenswrapper[4899]: I0123 08:53:28.968927 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:28Z","lastTransitionTime":"2026-01-23T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.071251 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.071291 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.071302 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.071324 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.071340 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.174605 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.174694 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.174714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.174738 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.174754 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.277503 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.277580 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.277602 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.277630 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.277651 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.380681 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.380827 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.380852 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.380882 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.380909 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.485148 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.485347 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.485784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.486365 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.486412 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.514482 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.514559 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:29 crc kubenswrapper[4899]: E0123 08:53:29.514611 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.514480 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.514739 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:29 crc kubenswrapper[4899]: E0123 08:53:29.514734 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:29 crc kubenswrapper[4899]: E0123 08:53:29.514920 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:29 crc kubenswrapper[4899]: E0123 08:53:29.515102 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.534628 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 17:17:31.906396776 +0000 UTC Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.589106 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.589171 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.589183 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.589198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.589259 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.691689 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.691760 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.691784 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.691809 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.691826 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.795789 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.795834 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.795845 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.795863 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.795873 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.898338 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.898395 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.898414 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.898452 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:29 crc kubenswrapper[4899]: I0123 08:53:29.898485 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:29Z","lastTransitionTime":"2026-01-23T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.001788 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.001852 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.001870 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.001905 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.001941 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.104925 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.104986 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.105009 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.105029 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.105041 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.208636 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.208694 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.208708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.208727 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.208737 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.311196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.311320 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.311335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.311354 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.311368 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.414622 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.414662 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.414675 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.414699 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.414722 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.517259 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.517334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.517352 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.517377 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.517395 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.535636 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 02:04:37.289007723 +0000 UTC Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.621252 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.621315 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.621334 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.621360 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.621373 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.725002 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.725075 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.725113 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.725146 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.725174 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.827681 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.827744 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.827758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.827778 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.827792 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.930472 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.930523 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.930537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.930556 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:30 crc kubenswrapper[4899]: I0123 08:53:30.930568 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:30Z","lastTransitionTime":"2026-01-23T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.033436 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.033499 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.033522 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.033551 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.033574 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.136019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.136083 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.136121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.136152 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.136174 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.239398 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.239461 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.239473 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.239497 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.239511 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.342360 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.342425 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.342441 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.342466 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.342483 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.444787 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.444852 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.444868 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.444893 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.444916 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.514078 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.514081 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:31 crc kubenswrapper[4899]: E0123 08:53:31.514273 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.514302 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:31 crc kubenswrapper[4899]: E0123 08:53:31.514480 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.514360 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:31 crc kubenswrapper[4899]: E0123 08:53:31.514830 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:31 crc kubenswrapper[4899]: E0123 08:53:31.514687 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.536675 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 01:19:28.80262679 +0000 UTC Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.548102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.548178 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.548202 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.548287 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.548310 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.650408 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.650471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.650479 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.650494 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.650503 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.752978 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.753014 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.753022 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.753035 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.753044 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.856154 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.856202 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.856240 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.856257 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.856270 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.958768 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.958820 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.958831 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.958844 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:31 crc kubenswrapper[4899]: I0123 08:53:31.958852 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:31Z","lastTransitionTime":"2026-01-23T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.062143 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.062193 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.062229 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.062249 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.062262 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.165578 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.165641 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.165654 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.165673 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.165691 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.268432 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.268481 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.268497 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.268517 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.268536 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.370502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.370537 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.370545 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.370558 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.370567 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.472621 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.472733 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.472743 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.472758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.472768 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.537238 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 23:45:28.08567197 +0000 UTC Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.575694 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.575787 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.575811 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.575842 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.575961 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.679164 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.679269 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.679290 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.679317 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.679334 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.782658 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.782718 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.782730 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.782750 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.782763 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.885110 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.885161 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.885172 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.885188 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.885200 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.988572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.988830 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.988884 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.988914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:32 crc kubenswrapper[4899]: I0123 08:53:32.988936 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:32Z","lastTransitionTime":"2026-01-23T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.091740 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.091820 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.091843 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.091873 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.091897 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.195405 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.195535 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.195565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.195597 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.195623 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.298639 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.298752 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.298816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.298842 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.298917 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.401490 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.401532 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.401545 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.401565 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.401578 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.503455 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.503505 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.503516 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.503534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.503550 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.514376 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.514446 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.514358 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:33 crc kubenswrapper[4899]: E0123 08:53:33.514518 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:33 crc kubenswrapper[4899]: E0123 08:53:33.514602 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.514683 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:33 crc kubenswrapper[4899]: E0123 08:53:33.514748 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:33 crc kubenswrapper[4899]: E0123 08:53:33.514897 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.538010 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:04:29.822618503 +0000 UTC Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.570048 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-lmzls" podStartSLOduration=76.570029041 podStartE2EDuration="1m16.570029041s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.568833691 +0000 UTC m=+100.435804098" watchObservedRunningTime="2026-01-23 08:53:33.570029041 +0000 UTC m=+100.436999438" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.606386 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h5gln" podStartSLOduration=76.606362877 podStartE2EDuration="1m16.606362877s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.59338536 +0000 UTC m=+100.460355747" watchObservedRunningTime="2026-01-23 08:53:33.606362877 +0000 UTC m=+100.473333264" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.607150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.607180 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.607192 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.607211 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.607238 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.655244 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.655185037 podStartE2EDuration="1m15.655185037s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.654996453 +0000 UTC m=+100.521966860" watchObservedRunningTime="2026-01-23 08:53:33.655185037 +0000 UTC m=+100.522155434" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.655447 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podStartSLOduration=76.655441444 podStartE2EDuration="1m16.655441444s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.615853776 +0000 UTC m=+100.482824253" watchObservedRunningTime="2026-01-23 08:53:33.655441444 +0000 UTC m=+100.522411841" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.691757 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.691738319 podStartE2EDuration="1m20.691738319s" podCreationTimestamp="2026-01-23 08:52:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.6795074 +0000 UTC m=+100.546477787" watchObservedRunningTime="2026-01-23 08:53:33.691738319 +0000 UTC m=+100.558708706" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.713070 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.713120 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.713137 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.713150 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.713160 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.771894 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=18.771876388 podStartE2EDuration="18.771876388s" podCreationTimestamp="2026-01-23 08:53:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.770090523 +0000 UTC m=+100.637060900" watchObservedRunningTime="2026-01-23 08:53:33.771876388 +0000 UTC m=+100.638846775" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.781424 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.781408569 podStartE2EDuration="48.781408569s" podCreationTimestamp="2026-01-23 08:52:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.7810734 +0000 UTC m=+100.648043787" watchObservedRunningTime="2026-01-23 08:53:33.781408569 +0000 UTC m=+100.648378956" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.815159 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.815228 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.815243 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.815260 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.815272 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.816437 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-nzqf7" podStartSLOduration=76.816426491 podStartE2EDuration="1m16.816426491s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.816296378 +0000 UTC m=+100.683266775" watchObservedRunningTime="2026-01-23 08:53:33.816426491 +0000 UTC m=+100.683396878" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.831435 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gbqr2" podStartSLOduration=76.831420239 podStartE2EDuration="1m16.831420239s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.830937857 +0000 UTC m=+100.697908254" watchObservedRunningTime="2026-01-23 08:53:33.831420239 +0000 UTC m=+100.698390626" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.843397 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khwdn" podStartSLOduration=75.84336894 podStartE2EDuration="1m15.84336894s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.8429705 +0000 UTC m=+100.709940887" watchObservedRunningTime="2026-01-23 08:53:33.84336894 +0000 UTC m=+100.710339327" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.868958 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.868930234 podStartE2EDuration="23.868930234s" podCreationTimestamp="2026-01-23 08:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:33.854985633 +0000 UTC m=+100.721956020" watchObservedRunningTime="2026-01-23 08:53:33.868930234 +0000 UTC m=+100.735900621" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.917612 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.917660 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.917669 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.917685 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:33 crc kubenswrapper[4899]: I0123 08:53:33.917695 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:33Z","lastTransitionTime":"2026-01-23T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.020020 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.020080 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.020096 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.020121 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.020139 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.122638 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.122687 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.122698 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.122714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.122728 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.225829 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.225896 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.225915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.225938 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.225953 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.327914 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.327952 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.327960 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.327977 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.327987 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.430461 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.430508 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.430523 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.430544 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.430558 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.533239 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.533293 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.533309 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.533329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.533344 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.538782 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 03:22:35.131489576 +0000 UTC Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.639314 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.639419 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.639450 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.639485 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.639522 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.742708 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.742781 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.742804 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.742825 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.742840 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.844781 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.844829 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.844841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.844858 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.844869 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.947777 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.947844 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.947860 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.947885 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:34 crc kubenswrapper[4899]: I0123 08:53:34.947902 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:34Z","lastTransitionTime":"2026-01-23T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.050071 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.050138 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.050157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.050181 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.050199 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.153102 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.153162 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.153178 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.153198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.153232 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.256381 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.256421 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.256430 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.256443 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.256452 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.358828 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.358915 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.358957 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.358989 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.359012 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.461758 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.461798 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.461806 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.461819 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.461828 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.514295 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.514362 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.514429 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:35 crc kubenswrapper[4899]: E0123 08:53:35.514431 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:35 crc kubenswrapper[4899]: E0123 08:53:35.514527 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.514573 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:35 crc kubenswrapper[4899]: E0123 08:53:35.514610 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:35 crc kubenswrapper[4899]: E0123 08:53:35.514686 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.539499 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 06:04:19.964084671 +0000 UTC Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.564068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.564105 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.564114 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.564126 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.564136 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.667094 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.667145 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.667154 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.667171 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.667180 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.769411 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.769463 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.769476 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.769494 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.769505 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.872896 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.872945 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.872959 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.872981 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.873007 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.976371 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.976442 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.976471 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.976505 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:35 crc kubenswrapper[4899]: I0123 08:53:35.976529 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:35Z","lastTransitionTime":"2026-01-23T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.038306 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:36 crc kubenswrapper[4899]: E0123 08:53:36.038573 4899 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:53:36 crc kubenswrapper[4899]: E0123 08:53:36.038684 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs podName:bb844e06-c862-406b-8ecb-36b682d2d58b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:40.038659566 +0000 UTC m=+166.905630013 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs") pod "network-metrics-daemon-kxcxw" (UID: "bb844e06-c862-406b-8ecb-36b682d2d58b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.080190 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.080328 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.080369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.080401 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.080428 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.183416 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.183475 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.183491 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.183513 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.183529 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.286233 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.286271 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.286279 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.286297 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.286314 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.388845 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.388904 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.388920 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.388944 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.388960 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.492554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.492639 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.492653 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.492675 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.492688 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.540672 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 18:57:30.126097326 +0000 UTC Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.595199 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.595253 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.595262 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.595276 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.595289 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.697997 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.698049 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.698061 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.698076 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.698089 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.800714 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.800797 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.800819 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.800846 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.800862 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.903078 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.903138 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.903156 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.903175 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:36 crc kubenswrapper[4899]: I0123 08:53:36.903186 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:36Z","lastTransitionTime":"2026-01-23T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.006124 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.006186 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.006198 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.006237 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.006249 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.109251 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.109304 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.109316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.109335 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.109346 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.212013 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.212062 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.212072 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.212087 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.212097 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.314816 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.314859 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.314868 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.314908 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.314919 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.419502 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.419576 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.419600 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.419630 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.419654 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.513693 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.513757 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.513796 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:37 crc kubenswrapper[4899]: E0123 08:53:37.513868 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.513939 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:37 crc kubenswrapper[4899]: E0123 08:53:37.514045 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:37 crc kubenswrapper[4899]: E0123 08:53:37.514113 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:37 crc kubenswrapper[4899]: E0123 08:53:37.514160 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.521550 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.521588 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.521600 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.521613 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.521622 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.541248 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 21:00:58.751724723 +0000 UTC Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.623988 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.624068 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.624077 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.624093 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.624105 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.726519 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.726560 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.726572 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.726587 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.726600 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.829232 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.829285 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.829295 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.829316 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.829329 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.932468 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.932522 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.932534 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.932554 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:37 crc kubenswrapper[4899]: I0123 08:53:37.932566 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:37Z","lastTransitionTime":"2026-01-23T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.035927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.035983 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.036027 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.036051 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.036063 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.139560 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.139614 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.139625 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.139646 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.139659 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.242930 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.243009 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.243023 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.243042 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.243059 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.345746 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.345792 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.345802 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.345825 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.345851 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.448974 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.449011 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.449019 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.449032 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.449040 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.542308 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 14:55:59.330650027 +0000 UTC Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.550927 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.550969 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.550980 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.550995 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.551006 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.653369 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.653418 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.653431 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.653447 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.653459 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.756157 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.756196 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.756207 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.756243 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.756255 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.858786 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.858841 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.858853 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.858874 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.858886 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.961244 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.961329 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.961347 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.961373 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:38 crc kubenswrapper[4899]: I0123 08:53:38.961389 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:38Z","lastTransitionTime":"2026-01-23T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.043486 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.043578 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.043602 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.043635 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.043657 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:39Z","lastTransitionTime":"2026-01-23T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.069963 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.070155 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.070193 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.070712 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.070817 4899 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-23T08:53:39Z","lastTransitionTime":"2026-01-23T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.101188 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584"] Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.101541 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.104335 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.104573 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.104749 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.106030 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.170189 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63fa2eca-54c4-45dd-88a3-992926bcd5f1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.170388 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fa2eca-54c4-45dd-88a3-992926bcd5f1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.170437 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.170491 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.170520 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63fa2eca-54c4-45dd-88a3-992926bcd5f1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271612 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271714 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271751 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63fa2eca-54c4-45dd-88a3-992926bcd5f1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271776 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271807 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63fa2eca-54c4-45dd-88a3-992926bcd5f1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.271842 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/63fa2eca-54c4-45dd-88a3-992926bcd5f1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.272050 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fa2eca-54c4-45dd-88a3-992926bcd5f1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.273197 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63fa2eca-54c4-45dd-88a3-992926bcd5f1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.285561 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fa2eca-54c4-45dd-88a3-992926bcd5f1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.299107 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63fa2eca-54c4-45dd-88a3-992926bcd5f1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh584\" (UID: \"63fa2eca-54c4-45dd-88a3-992926bcd5f1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.415445 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" Jan 23 08:53:39 crc kubenswrapper[4899]: W0123 08:53:39.431491 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fa2eca_54c4_45dd_88a3_992926bcd5f1.slice/crio-51ba89a0b1a5b81b87394bb59d80f1f6a9a783a1c42f804c45597c2af9cafa30 WatchSource:0}: Error finding container 51ba89a0b1a5b81b87394bb59d80f1f6a9a783a1c42f804c45597c2af9cafa30: Status 404 returned error can't find the container with id 51ba89a0b1a5b81b87394bb59d80f1f6a9a783a1c42f804c45597c2af9cafa30 Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.514099 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:39 crc kubenswrapper[4899]: E0123 08:53:39.514312 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.514624 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:39 crc kubenswrapper[4899]: E0123 08:53:39.514734 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.515311 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.515867 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:39 crc kubenswrapper[4899]: E0123 08:53:39.516080 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.516251 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:53:39 crc kubenswrapper[4899]: E0123 08:53:39.516337 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:39 crc kubenswrapper[4899]: E0123 08:53:39.517256 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.543318 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 05:49:10.20665436 +0000 UTC Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.543377 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 23 08:53:39 crc kubenswrapper[4899]: I0123 08:53:39.551708 4899 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 23 08:53:40 crc kubenswrapper[4899]: I0123 08:53:40.063135 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" event={"ID":"63fa2eca-54c4-45dd-88a3-992926bcd5f1","Type":"ContainerStarted","Data":"0f6f877d63241a80dc40081c3f9e07843bde45c9c10d097ed747bb8a9d2be40a"} Jan 23 08:53:40 crc kubenswrapper[4899]: I0123 08:53:40.063276 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" event={"ID":"63fa2eca-54c4-45dd-88a3-992926bcd5f1","Type":"ContainerStarted","Data":"51ba89a0b1a5b81b87394bb59d80f1f6a9a783a1c42f804c45597c2af9cafa30"} Jan 23 08:53:41 crc kubenswrapper[4899]: I0123 08:53:41.514192 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:41 crc kubenswrapper[4899]: I0123 08:53:41.514284 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:41 crc kubenswrapper[4899]: I0123 08:53:41.514398 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:41 crc kubenswrapper[4899]: I0123 08:53:41.514598 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:41 crc kubenswrapper[4899]: E0123 08:53:41.514582 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:41 crc kubenswrapper[4899]: E0123 08:53:41.514777 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:41 crc kubenswrapper[4899]: E0123 08:53:41.514938 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:41 crc kubenswrapper[4899]: E0123 08:53:41.515051 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:43 crc kubenswrapper[4899]: I0123 08:53:43.514422 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:43 crc kubenswrapper[4899]: I0123 08:53:43.514596 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:43 crc kubenswrapper[4899]: E0123 08:53:43.516346 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:43 crc kubenswrapper[4899]: I0123 08:53:43.516429 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:43 crc kubenswrapper[4899]: I0123 08:53:43.516677 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:43 crc kubenswrapper[4899]: E0123 08:53:43.516691 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:43 crc kubenswrapper[4899]: E0123 08:53:43.516898 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:43 crc kubenswrapper[4899]: E0123 08:53:43.516989 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:45 crc kubenswrapper[4899]: I0123 08:53:45.514489 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:45 crc kubenswrapper[4899]: I0123 08:53:45.514568 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:45 crc kubenswrapper[4899]: E0123 08:53:45.514981 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:45 crc kubenswrapper[4899]: I0123 08:53:45.514799 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:45 crc kubenswrapper[4899]: I0123 08:53:45.514629 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:45 crc kubenswrapper[4899]: E0123 08:53:45.515092 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:45 crc kubenswrapper[4899]: E0123 08:53:45.515274 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:45 crc kubenswrapper[4899]: E0123 08:53:45.515392 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:47 crc kubenswrapper[4899]: I0123 08:53:47.513974 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:47 crc kubenswrapper[4899]: I0123 08:53:47.514065 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:47 crc kubenswrapper[4899]: I0123 08:53:47.514077 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:47 crc kubenswrapper[4899]: I0123 08:53:47.514015 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:47 crc kubenswrapper[4899]: E0123 08:53:47.514245 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:47 crc kubenswrapper[4899]: E0123 08:53:47.514384 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:47 crc kubenswrapper[4899]: E0123 08:53:47.514540 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:47 crc kubenswrapper[4899]: E0123 08:53:47.514747 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:49 crc kubenswrapper[4899]: I0123 08:53:49.514418 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:49 crc kubenswrapper[4899]: I0123 08:53:49.514448 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:49 crc kubenswrapper[4899]: E0123 08:53:49.514542 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:49 crc kubenswrapper[4899]: I0123 08:53:49.514583 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:49 crc kubenswrapper[4899]: I0123 08:53:49.514639 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:49 crc kubenswrapper[4899]: E0123 08:53:49.514761 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:49 crc kubenswrapper[4899]: E0123 08:53:49.515048 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:49 crc kubenswrapper[4899]: E0123 08:53:49.515145 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.100229 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/1.log" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.100630 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/0.log" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.100666 4899 generic.go:334] "Generic (PLEG): container finished" podID="60a84ebf-a188-4eee-8a9f-e02ac7adb8c8" containerID="cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697" exitCode=1 Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.100696 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerDied","Data":"cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697"} Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.100729 4899 scope.go:117] "RemoveContainer" containerID="e0fb7ac9e65a75799a3b0f39b7850d96267ef0174a94192c04d84dc1b0c2acb6" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.101519 4899 scope.go:117] "RemoveContainer" containerID="cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.101792 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h5gln_openshift-multus(60a84ebf-a188-4eee-8a9f-e02ac7adb8c8)\"" pod="openshift-multus/multus-h5gln" podUID="60a84ebf-a188-4eee-8a9f-e02ac7adb8c8" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.126189 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh584" podStartSLOduration=94.126167443 podStartE2EDuration="1m34.126167443s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:53:40.085021651 +0000 UTC m=+106.951992048" watchObservedRunningTime="2026-01-23 08:53:51.126167443 +0000 UTC m=+117.993137820" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.514025 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.514169 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.514311 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.514288 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.514463 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.514610 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.514049 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.515462 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:51 crc kubenswrapper[4899]: I0123 08:53:51.515917 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:53:51 crc kubenswrapper[4899]: E0123 08:53:51.516506 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qmqs_openshift-ovn-kubernetes(790411d3-0df3-4cc3-bf33-835dab4c2cb7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" Jan 23 08:53:52 crc kubenswrapper[4899]: I0123 08:53:52.105037 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/1.log" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.502900 4899 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 23 08:53:53 crc kubenswrapper[4899]: I0123 08:53:53.569883 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.571288 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:53 crc kubenswrapper[4899]: I0123 08:53:53.571334 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:53 crc kubenswrapper[4899]: I0123 08:53:53.571375 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.571456 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.571533 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:53 crc kubenswrapper[4899]: I0123 08:53:53.571770 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.572494 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:53 crc kubenswrapper[4899]: E0123 08:53:53.670679 4899 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 23 08:53:55 crc kubenswrapper[4899]: I0123 08:53:55.514532 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:55 crc kubenswrapper[4899]: I0123 08:53:55.514583 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:55 crc kubenswrapper[4899]: I0123 08:53:55.514678 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:55 crc kubenswrapper[4899]: E0123 08:53:55.514686 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:55 crc kubenswrapper[4899]: E0123 08:53:55.514829 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:55 crc kubenswrapper[4899]: E0123 08:53:55.514973 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:55 crc kubenswrapper[4899]: I0123 08:53:55.515400 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:55 crc kubenswrapper[4899]: E0123 08:53:55.515477 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:57 crc kubenswrapper[4899]: I0123 08:53:57.513966 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:57 crc kubenswrapper[4899]: I0123 08:53:57.514011 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:57 crc kubenswrapper[4899]: I0123 08:53:57.514063 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:57 crc kubenswrapper[4899]: E0123 08:53:57.514166 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:53:57 crc kubenswrapper[4899]: I0123 08:53:57.514318 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:57 crc kubenswrapper[4899]: E0123 08:53:57.514392 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:57 crc kubenswrapper[4899]: E0123 08:53:57.514325 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:57 crc kubenswrapper[4899]: E0123 08:53:57.514515 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:58 crc kubenswrapper[4899]: E0123 08:53:58.671991 4899 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 23 08:53:59 crc kubenswrapper[4899]: I0123 08:53:59.514024 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:53:59 crc kubenswrapper[4899]: I0123 08:53:59.514077 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:53:59 crc kubenswrapper[4899]: I0123 08:53:59.514187 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:53:59 crc kubenswrapper[4899]: I0123 08:53:59.514227 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:53:59 crc kubenswrapper[4899]: E0123 08:53:59.514308 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:53:59 crc kubenswrapper[4899]: E0123 08:53:59.514411 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:53:59 crc kubenswrapper[4899]: E0123 08:53:59.514497 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:53:59 crc kubenswrapper[4899]: E0123 08:53:59.514556 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:54:01 crc kubenswrapper[4899]: I0123 08:54:01.513938 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:01 crc kubenswrapper[4899]: I0123 08:54:01.513982 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:01 crc kubenswrapper[4899]: I0123 08:54:01.514032 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:01 crc kubenswrapper[4899]: E0123 08:54:01.514112 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:54:01 crc kubenswrapper[4899]: I0123 08:54:01.514130 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:01 crc kubenswrapper[4899]: E0123 08:54:01.514312 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:54:01 crc kubenswrapper[4899]: E0123 08:54:01.514300 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:54:01 crc kubenswrapper[4899]: E0123 08:54:01.514360 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:54:03 crc kubenswrapper[4899]: I0123 08:54:03.516561 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:03 crc kubenswrapper[4899]: I0123 08:54:03.516653 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:03 crc kubenswrapper[4899]: I0123 08:54:03.516723 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:03 crc kubenswrapper[4899]: E0123 08:54:03.516853 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:54:03 crc kubenswrapper[4899]: I0123 08:54:03.516897 4899 scope.go:117] "RemoveContainer" containerID="cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697" Jan 23 08:54:03 crc kubenswrapper[4899]: I0123 08:54:03.516984 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:03 crc kubenswrapper[4899]: E0123 08:54:03.517121 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:54:03 crc kubenswrapper[4899]: E0123 08:54:03.517534 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:54:03 crc kubenswrapper[4899]: E0123 08:54:03.517658 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:54:03 crc kubenswrapper[4899]: E0123 08:54:03.672470 4899 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 23 08:54:04 crc kubenswrapper[4899]: I0123 08:54:04.147545 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/1.log" Jan 23 08:54:04 crc kubenswrapper[4899]: I0123 08:54:04.147601 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerStarted","Data":"f9014f9495c81d445c7b91ebf5490a70e7db4b2d956003d1991b6e84ed06afce"} Jan 23 08:54:04 crc kubenswrapper[4899]: I0123 08:54:04.514636 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.152613 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/3.log" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.154910 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerStarted","Data":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.155269 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.179012 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podStartSLOduration=108.178993865 podStartE2EDuration="1m48.178993865s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:05.178611136 +0000 UTC m=+132.045581523" watchObservedRunningTime="2026-01-23 08:54:05.178993865 +0000 UTC m=+132.045964252" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.312694 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kxcxw"] Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.312801 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:05 crc kubenswrapper[4899]: E0123 08:54:05.312885 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.514436 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:05 crc kubenswrapper[4899]: E0123 08:54:05.514556 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.514737 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:05 crc kubenswrapper[4899]: E0123 08:54:05.514788 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:54:05 crc kubenswrapper[4899]: I0123 08:54:05.514984 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:05 crc kubenswrapper[4899]: E0123 08:54:05.515042 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:54:07 crc kubenswrapper[4899]: I0123 08:54:07.513606 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:07 crc kubenswrapper[4899]: I0123 08:54:07.513642 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:07 crc kubenswrapper[4899]: I0123 08:54:07.513720 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:07 crc kubenswrapper[4899]: E0123 08:54:07.513757 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 23 08:54:07 crc kubenswrapper[4899]: E0123 08:54:07.513879 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxcxw" podUID="bb844e06-c862-406b-8ecb-36b682d2d58b" Jan 23 08:54:07 crc kubenswrapper[4899]: I0123 08:54:07.513951 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:07 crc kubenswrapper[4899]: E0123 08:54:07.513977 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 23 08:54:07 crc kubenswrapper[4899]: E0123 08:54:07.514039 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.514488 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.514521 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.514583 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.514750 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.516639 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.516640 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.517102 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.517449 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.518633 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.519268 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.640509 4899 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.676236 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.676637 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k4cjk"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.677050 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.677396 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.679897 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.682184 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.682560 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.682906 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.683495 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.683554 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.683802 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.684007 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.684445 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.684652 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.684824 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.690465 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.690708 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.694368 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.694793 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.695091 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.695212 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhxvh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.695283 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.695338 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.695534 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.705026 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.705692 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.712372 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.715805 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-r6525"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.715876 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.716152 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqzln"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.716715 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.716786 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4hxzp"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.717114 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.717139 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.718269 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-n7s8j"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.718730 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.718990 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.726550 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.727097 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.727533 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.729738 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.731588 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.732899 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.733554 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.759727 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ctjz9"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.760129 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.764874 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.764916 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.784427 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.784529 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.784769 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.803528 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.803805 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.804041 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.804493 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tbwm2"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.804770 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.804993 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.805184 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.805406 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.807189 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.807364 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.807506 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.807671 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.807840 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808021 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808161 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808338 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808411 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808529 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808581 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808726 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808849 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.808909 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809028 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809332 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809416 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809451 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809574 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809629 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809714 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809856 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809919 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.809964 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810063 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810129 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810409 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810420 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810482 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810497 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810552 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810559 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810627 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810639 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810710 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810774 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810810 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810839 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810909 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810949 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810976 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.810776 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.811098 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.811349 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.811418 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.811802 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.812088 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.834064 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.839425 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.840324 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.840970 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.841059 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.841111 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.849594 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852229 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852422 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk978\" (UniqueName: \"kubernetes.io/projected/986b43bc-5ff7-486a-8539-f7568bc468f1-kube-api-access-mk978\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852450 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852523 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-serving-cert\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852539 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-service-ca\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852554 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-oauth-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852568 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdj6h\" (UniqueName: \"kubernetes.io/projected/0be4b82f-6b52-486a-8f68-88a6cd3a8317-kube-api-access-qdj6h\") pod \"downloads-7954f5f757-r6525\" (UID: \"0be4b82f-6b52-486a-8f68-88a6cd3a8317\") " pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852659 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-oauth-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852680 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk6d4\" (UniqueName: \"kubernetes.io/projected/e0e33477-d346-481d-8e80-bb90445a0eb2-kube-api-access-zk6d4\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852696 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-client\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852967 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-node-pullsecrets\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.852995 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853025 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-audit\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853041 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-audit-dir\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853065 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853104 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a7de32e-498b-46e2-9de3-7313c34eab26-serving-cert\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853125 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853160 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp7wc\" (UniqueName: \"kubernetes.io/projected/3709c439-2199-443d-8e3d-250b85dfdf2a-kube-api-access-mp7wc\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853179 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-trusted-ca-bundle\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853194 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-policies\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853383 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-encryption-config\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853444 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853472 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433100b8-a3d5-4f8a-8e88-57026c742941-serving-cert\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853491 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftnwt\" (UniqueName: \"kubernetes.io/projected/433100b8-a3d5-4f8a-8e88-57026c742941-kube-api-access-ftnwt\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853506 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fmjn\" (UniqueName: \"kubernetes.io/projected/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-kube-api-access-4fmjn\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853529 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853549 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-serving-cert\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853571 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhqnr\" (UniqueName: \"kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853587 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853611 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4sj5\" (UniqueName: \"kubernetes.io/projected/4a7de32e-498b-46e2-9de3-7313c34eab26-kube-api-access-b4sj5\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853633 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-serving-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853649 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-encryption-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853671 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f626d106-0870-45f8-ab43-31d92627d068-machine-approver-tls\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853692 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853711 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853744 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-trusted-ca\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853760 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96n9w\" (UniqueName: \"kubernetes.io/projected/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-kube-api-access-96n9w\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853775 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjjp9\" (UniqueName: \"kubernetes.io/projected/f626d106-0870-45f8-ab43-31d92627d068-kube-api-access-bjjp9\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853793 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853806 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853822 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-config\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853855 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-auth-proxy-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853875 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853892 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853918 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-client\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853939 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-image-import-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853954 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853971 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3709c439-2199-443d-8e3d-250b85dfdf2a-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853984 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3709c439-2199-443d-8e3d-250b85dfdf2a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.853999 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854016 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854030 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6wcm\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-kube-api-access-h6wcm\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854044 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854081 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-dir\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854097 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-config\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854110 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bs5w\" (UniqueName: \"kubernetes.io/projected/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-kube-api-access-5bs5w\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854140 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854647 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854921 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.854689 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k4cjk"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.855232 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.855462 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.855621 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.855765 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.855886 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h5tlv"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.856308 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.856636 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.857592 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858248 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858339 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858411 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858520 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858602 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.858710 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.859668 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-clf44"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.860391 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.862047 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.864201 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.864282 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.864698 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-hszft"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.864978 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.865042 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.865063 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.865784 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.865830 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.867291 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.867640 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.868762 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.869099 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.869921 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.870349 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.870731 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.870861 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.870973 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.871248 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.871352 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.871454 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.871545 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.872136 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ctjz9"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.873101 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.873566 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.874240 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.878713 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.879151 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.881547 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4hxzp"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.891249 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.893166 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.897473 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhxvh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.898901 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.898906 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.900154 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.902020 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.902307 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.908388 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.908727 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.908815 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.908841 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.909881 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.911634 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6jqvj"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.912464 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.912558 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.913144 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.914318 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.914669 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.915629 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzm66"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.916268 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.917695 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.918577 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqzln"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.919072 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.919808 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.920264 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.921427 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.921987 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.922248 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.922646 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.923302 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjw59"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.923874 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.924765 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.925374 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.925973 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2f5n6"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.926395 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.927154 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-n7s8j"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.928556 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.931685 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.932795 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.933817 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r6525"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.934965 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-clf44"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.936245 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.937634 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.938452 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tbwm2"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.939415 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.939626 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.940837 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cbtmt"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.942020 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-flx89"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.942436 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.942434 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.944161 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.944190 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.946825 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.946883 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6jqvj"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.948445 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.949474 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h5tlv"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.950478 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.951489 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.952479 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.953633 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzm66"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954511 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-client\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954551 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad27e9b-b533-485b-9ec1-ac004ed76611-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954603 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-image-import-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954625 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954643 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3709c439-2199-443d-8e3d-250b85dfdf2a-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954658 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3709c439-2199-443d-8e3d-250b85dfdf2a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954676 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954694 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-service-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954711 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954734 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954750 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6wcm\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-kube-api-access-h6wcm\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954766 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bs5w\" (UniqueName: \"kubernetes.io/projected/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-kube-api-access-5bs5w\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954789 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954806 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-dir\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954822 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-config\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954841 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954858 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccz5d\" (UniqueName: \"kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954875 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk978\" (UniqueName: \"kubernetes.io/projected/986b43bc-5ff7-486a-8539-f7568bc468f1-kube-api-access-mk978\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954892 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954908 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954928 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954944 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-serving-cert\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954958 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-service-ca\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954975 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-oauth-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.954990 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdj6h\" (UniqueName: \"kubernetes.io/projected/0be4b82f-6b52-486a-8f68-88a6cd3a8317-kube-api-access-qdj6h\") pod \"downloads-7954f5f757-r6525\" (UID: \"0be4b82f-6b52-486a-8f68-88a6cd3a8317\") " pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955006 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955023 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-client\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955038 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-oauth-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955054 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk6d4\" (UniqueName: \"kubernetes.io/projected/e0e33477-d346-481d-8e80-bb90445a0eb2-kube-api-access-zk6d4\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955071 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955089 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955103 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955121 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955136 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-node-pullsecrets\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955152 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955169 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955185 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955200 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6p5h\" (UniqueName: \"kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955237 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-audit\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955259 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-audit-dir\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955283 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955304 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-config\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955324 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955343 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-serving-cert\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955378 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a7de32e-498b-46e2-9de3-7313c34eab26-serving-cert\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955390 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955399 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955415 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955430 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955446 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955469 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955484 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv74k\" (UniqueName: \"kubernetes.io/projected/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-kube-api-access-lv74k\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955500 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp7wc\" (UniqueName: \"kubernetes.io/projected/3709c439-2199-443d-8e3d-250b85dfdf2a-kube-api-access-mp7wc\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955517 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-encryption-config\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955532 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-trusted-ca-bundle\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955546 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-policies\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955563 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955578 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzjgb\" (UniqueName: \"kubernetes.io/projected/9ad27e9b-b533-485b-9ec1-ac004ed76611-kube-api-access-hzjgb\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955582 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-image-import-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955594 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433100b8-a3d5-4f8a-8e88-57026c742941-serving-cert\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955612 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftnwt\" (UniqueName: \"kubernetes.io/projected/433100b8-a3d5-4f8a-8e88-57026c742941-kube-api-access-ftnwt\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955622 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955628 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fmjn\" (UniqueName: \"kubernetes.io/projected/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-kube-api-access-4fmjn\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955648 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955663 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955675 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-node-pullsecrets\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955683 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955698 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-etcd-client\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955714 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955731 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhqnr\" (UniqueName: \"kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955746 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955760 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-serving-cert\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955778 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-images\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955795 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955806 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3709c439-2199-443d-8e3d-250b85dfdf2a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955819 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955811 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-serving-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955872 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-encryption-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955898 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4sj5\" (UniqueName: \"kubernetes.io/projected/4a7de32e-498b-46e2-9de3-7313c34eab26-kube-api-access-b4sj5\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955921 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955946 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f626d106-0870-45f8-ab43-31d92627d068-machine-approver-tls\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955967 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-config\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.955993 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956016 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956038 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnp29\" (UniqueName: \"kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956070 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjjp9\" (UniqueName: \"kubernetes.io/projected/f626d106-0870-45f8-ab43-31d92627d068-kube-api-access-bjjp9\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956093 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956114 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956140 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-config\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956162 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-trusted-ca\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956183 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96n9w\" (UniqueName: \"kubernetes.io/projected/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-kube-api-access-96n9w\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956228 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pq7b\" (UniqueName: \"kubernetes.io/projected/2376228d-4a51-459a-ad64-d152e53bf109-kube-api-access-9pq7b\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956256 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-auth-proxy-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956278 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956303 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956324 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956826 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956844 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.957237 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.956231 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-serving-ca\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.958555 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-audit\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.958883 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-config\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.958894 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-auth-proxy-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.959403 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.959430 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.959485 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.959915 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.960201 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-trusted-ca\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.960548 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3709c439-2199-443d-8e3d-250b85dfdf2a-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.960868 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-oauth-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.961497 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-service-ca\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.961732 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-etcd-client\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.962063 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-policies\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.962159 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/986b43bc-5ff7-486a-8539-f7568bc468f1-audit-dir\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.962294 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f626d106-0870-45f8-ab43-31d92627d068-config\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.963117 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.963631 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f626d106-0870-45f8-ab43-31d92627d068-machine-approver-tls\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.963764 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.964761 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/433100b8-a3d5-4f8a-8e88-57026c742941-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965048 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433100b8-a3d5-4f8a-8e88-57026c742941-serving-cert\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965110 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-serving-cert\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965332 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-serving-cert\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965354 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-trusted-ca-bundle\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965563 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.965822 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e33477-d346-481d-8e80-bb90445a0eb2-audit-dir\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.966726 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.966873 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-encryption-config\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.966923 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.966963 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.967519 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-flx89"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.967606 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7de32e-498b-46e2-9de3-7313c34eab26-config\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.967716 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a7de32e-498b-46e2-9de3-7313c34eab26-serving-cert\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.968374 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-serving-cert\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.970867 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/986b43bc-5ff7-486a-8539-f7568bc468f1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.970910 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.974409 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-console-oauth-config\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.974014 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.974524 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.974573 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cbtmt"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.974612 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/986b43bc-5ff7-486a-8539-f7568bc468f1-encryption-config\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.975564 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0e33477-d346-481d-8e80-bb90445a0eb2-etcd-client\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.976013 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hgqbd"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.976366 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.976845 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.977791 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjw59"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.979108 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hgqbd"] Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.986978 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 23 08:54:09 crc kubenswrapper[4899]: I0123 08:54:09.999460 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.019302 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.044494 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057085 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-config\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057314 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnp29\" (UniqueName: \"kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057462 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pq7b\" (UniqueName: \"kubernetes.io/projected/2376228d-4a51-459a-ad64-d152e53bf109-kube-api-access-9pq7b\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057571 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057669 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad27e9b-b533-485b-9ec1-ac004ed76611-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057777 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-service-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057909 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058010 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccz5d\" (UniqueName: \"kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058115 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058273 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.057786 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-config\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058380 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058437 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058489 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058528 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058552 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058573 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058590 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058611 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058628 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-config\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058642 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6p5h\" (UniqueName: \"kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058660 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-serving-cert\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058681 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058707 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058721 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058748 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058762 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv74k\" (UniqueName: \"kubernetes.io/projected/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-kube-api-access-lv74k\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058792 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058821 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058817 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058836 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzjgb\" (UniqueName: \"kubernetes.io/projected/9ad27e9b-b533-485b-9ec1-ac004ed76611-kube-api-access-hzjgb\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058905 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-etcd-client\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.058941 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.059011 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-images\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.059062 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.059083 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.060462 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-images\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.061305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-service-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.061862 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.061886 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.062543 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.062651 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.062726 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.062753 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.062801 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.063002 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.063132 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.063529 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.063534 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.065009 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.065171 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.065291 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.065618 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.069387 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-etcd-ca\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.069413 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-etcd-client\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.069756 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2376228d-4a51-459a-ad64-d152e53bf109-config\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.071100 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2376228d-4a51-459a-ad64-d152e53bf109-serving-cert\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.079309 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.098639 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.119903 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.139846 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.158733 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.179114 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.200100 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.220029 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.238695 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.259014 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.279261 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.298972 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.319785 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.339418 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.359079 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.379946 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.400492 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.419949 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.439048 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.459618 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.479551 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.498824 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.518758 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.538869 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.558856 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.579519 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.598597 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.619611 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.640860 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.659637 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.679697 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.700019 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.720274 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.739618 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.759879 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.781000 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.799971 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.819146 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.838929 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.863533 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.879859 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.899311 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.918034 4899 request.go:700] Waited for 1.007989506s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackage-server-manager-serving-cert&limit=500&resourceVersion=0 Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.919315 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.933566 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad27e9b-b533-485b-9ec1-ac004ed76611-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.938501 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.960510 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.980204 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 23 08:54:10 crc kubenswrapper[4899]: I0123 08:54:10.999744 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.020769 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.040160 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 23 08:54:11 crc kubenswrapper[4899]: E0123 08:54:11.059202 4899 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Jan 23 08:54:11 crc kubenswrapper[4899]: E0123 08:54:11.059333 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics podName:7d51b363-c51f-4ccd-b7ca-77ad850024dc nodeName:}" failed. No retries permitted until 2026-01-23 08:54:11.559303512 +0000 UTC m=+138.426273909 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics") pod "marketplace-operator-79b997595-wcqlx" (UID: "7d51b363-c51f-4ccd-b7ca-77ad850024dc") : failed to sync secret cache: timed out waiting for the condition Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.059340 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 23 08:54:11 crc kubenswrapper[4899]: E0123 08:54:11.059376 4899 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Jan 23 08:54:11 crc kubenswrapper[4899]: E0123 08:54:11.059489 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca podName:7d51b363-c51f-4ccd-b7ca-77ad850024dc nodeName:}" failed. No retries permitted until 2026-01-23 08:54:11.559459346 +0000 UTC m=+138.426429773 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca") pod "marketplace-operator-79b997595-wcqlx" (UID: "7d51b363-c51f-4ccd-b7ca-77ad850024dc") : failed to sync configmap cache: timed out waiting for the condition Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.079518 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.098983 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.119839 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.139437 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.159235 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.179015 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.199329 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.228482 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.238843 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.259173 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.279438 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.299028 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.319762 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.339655 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.359987 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.379561 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.399305 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.419068 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.440248 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.460888 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.480598 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.499843 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.520796 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.540810 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.559449 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.579201 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.582275 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.582349 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.583769 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.585157 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.600004 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.620111 4899 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.639128 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.659867 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.713607 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96n9w\" (UniqueName: \"kubernetes.io/projected/ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3-kube-api-access-96n9w\") pod \"openshift-apiserver-operator-796bbdcf4f-555np\" (UID: \"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.738037 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.758432 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk6d4\" (UniqueName: \"kubernetes.io/projected/e0e33477-d346-481d-8e80-bb90445a0eb2-kube-api-access-zk6d4\") pod \"apiserver-7bbb656c7d-5bwgh\" (UID: \"e0e33477-d346-481d-8e80-bb90445a0eb2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.771816 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjjp9\" (UniqueName: \"kubernetes.io/projected/f626d106-0870-45f8-ab43-31d92627d068-kube-api-access-bjjp9\") pod \"machine-approver-56656f9798-62gn5\" (UID: \"f626d106-0870-45f8-ab43-31d92627d068\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.791583 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdj6h\" (UniqueName: \"kubernetes.io/projected/0be4b82f-6b52-486a-8f68-88a6cd3a8317-kube-api-access-qdj6h\") pod \"downloads-7954f5f757-r6525\" (UID: \"0be4b82f-6b52-486a-8f68-88a6cd3a8317\") " pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.812695 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fmjn\" (UniqueName: \"kubernetes.io/projected/27183ff6-88f1-42ef-9d0c-bbd4da6d2441-kube-api-access-4fmjn\") pod \"console-f9d7485db-n7s8j\" (UID: \"27183ff6-88f1-42ef-9d0c-bbd4da6d2441\") " pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.833398 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk978\" (UniqueName: \"kubernetes.io/projected/986b43bc-5ff7-486a-8539-f7568bc468f1-kube-api-access-mk978\") pod \"apiserver-76f77b778f-k4cjk\" (UID: \"986b43bc-5ff7-486a-8539-f7568bc468f1\") " pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.847122 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.852885 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftnwt\" (UniqueName: \"kubernetes.io/projected/433100b8-a3d5-4f8a-8e88-57026c742941-kube-api-access-ftnwt\") pod \"authentication-operator-69f744f599-hhxvh\" (UID: \"433100b8-a3d5-4f8a-8e88-57026c742941\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.856002 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.865424 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.872817 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.875092 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhqnr\" (UniqueName: \"kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr\") pod \"controller-manager-879f6c89f-vfz69\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.902468 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.932583 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.936124 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp7wc\" (UniqueName: \"kubernetes.io/projected/3709c439-2199-443d-8e3d-250b85dfdf2a-kube-api-access-mp7wc\") pod \"openshift-config-operator-7777fb866f-zqzln\" (UID: \"3709c439-2199-443d-8e3d-250b85dfdf2a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.937454 4899 request.go:700] Waited for 1.970810482s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.941601 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bs5w\" (UniqueName: \"kubernetes.io/projected/ec02f16b-18e4-4bf8-a921-2104a1f7ffff-kube-api-access-5bs5w\") pod \"cluster-samples-operator-665b6dd947-lnnll\" (UID: \"ec02f16b-18e4-4bf8-a921-2104a1f7ffff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.941968 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6wcm\" (UniqueName: \"kubernetes.io/projected/b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2-kube-api-access-h6wcm\") pod \"cluster-image-registry-operator-dc59b4c8b-bwh2k\" (UID: \"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.955724 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4sj5\" (UniqueName: \"kubernetes.io/projected/4a7de32e-498b-46e2-9de3-7313c34eab26-kube-api-access-b4sj5\") pod \"console-operator-58897d9998-4hxzp\" (UID: \"4a7de32e-498b-46e2-9de3-7313c34eab26\") " pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.955928 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.959663 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.966880 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.981998 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 23 08:54:11 crc kubenswrapper[4899]: I0123 08:54:11.999795 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.043750 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pq7b\" (UniqueName: \"kubernetes.io/projected/2376228d-4a51-459a-ad64-d152e53bf109-kube-api-access-9pq7b\") pod \"etcd-operator-b45778765-h5tlv\" (UID: \"2376228d-4a51-459a-ad64-d152e53bf109\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.054872 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnp29\" (UniqueName: \"kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29\") pod \"oauth-openshift-558db77b4-ctjz9\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.076645 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccz5d\" (UniqueName: \"kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d\") pod \"route-controller-manager-6576b87f9c-2xtw6\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.110086 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzjgb\" (UniqueName: \"kubernetes.io/projected/9ad27e9b-b533-485b-9ec1-ac004ed76611-kube-api-access-hzjgb\") pod \"package-server-manager-789f6589d5-ddgpg\" (UID: \"9ad27e9b-b533-485b-9ec1-ac004ed76611\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.119634 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6p5h\" (UniqueName: \"kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h\") pod \"marketplace-operator-79b997595-wcqlx\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.125368 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.137372 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.138056 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv74k\" (UniqueName: \"kubernetes.io/projected/926c5bb9-7f91-4ea9-860c-694ec9c5c4db-kube-api-access-lv74k\") pod \"machine-api-operator-5694c8668f-tbwm2\" (UID: \"926c5bb9-7f91-4ea9-860c-694ec9c5c4db\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.157640 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.185531 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" event={"ID":"f626d106-0870-45f8-ab43-31d92627d068","Type":"ContainerStarted","Data":"124d49cacb205832899a9478461a85edf9f4bfb3233473b5f0a0f17c68fb7453"} Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.185761 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192653 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8cd6278-992f-4d7b-81ce-9f444bac3730-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192716 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192744 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-srv-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192770 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/630c4c0a-8622-4322-b217-49b25827730e-signing-cabundle\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192795 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwzpv\" (UniqueName: \"kubernetes.io/projected/ead3bf87-613f-4624-9eaa-29fd11f02cfd-kube-api-access-hwzpv\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192833 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/984dfa93-96fc-401c-a7ad-61d60a9f77f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192856 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e413497-e6a2-4e02-afb7-12be8584bb54-proxy-tls\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192878 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9h4p\" (UniqueName: \"kubernetes.io/projected/db7591f0-3d80-418a-b4f3-850134650dea-kube-api-access-h9h4p\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192899 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192923 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-webhook-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192947 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-certs\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.192971 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk555\" (UniqueName: \"kubernetes.io/projected/d90c2cf7-520f-45f3-9c6d-e3970848f073-kube-api-access-qk555\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193004 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984dfa93-96fc-401c-a7ad-61d60a9f77f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193037 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwdkv\" (UniqueName: \"kubernetes.io/projected/5e413497-e6a2-4e02-afb7-12be8584bb54-kube-api-access-bwdkv\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193072 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c6affa4-e748-447d-85d9-78c91b480940-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193096 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ea85a4b-f44f-4500-94d8-99c116ab4e43-service-ca-bundle\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193122 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db7591f0-3d80-418a-b4f3-850134650dea-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193145 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193168 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad2193d1-bbed-4942-a2e5-40246e3c67de-trusted-ca\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.193934 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:12.693919858 +0000 UTC m=+139.560890245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.193207 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/191860f0-d71a-44e8-b5ca-5eab39d390c0-cert\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.197274 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.197329 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c6affa4-e748-447d-85d9-78c91b480940-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.197469 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.198035 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-csi-data-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.198067 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-mountpoint-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.198933 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-metrics-certs\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.198953 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r56hl\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-kube-api-access-r56hl\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.198984 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-proxy-tls\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199010 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpm6x\" (UniqueName: \"kubernetes.io/projected/6ea85a4b-f44f-4500-94d8-99c116ab4e43-kube-api-access-tpm6x\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199030 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkdk6\" (UniqueName: \"kubernetes.io/projected/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-kube-api-access-jkdk6\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199053 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjmp7\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199073 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199114 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199137 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-889vk\" (UniqueName: \"kubernetes.io/projected/8957b168-510a-4813-9036-900e09d378f5-kube-api-access-889vk\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199159 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-registration-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199189 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-default-certificate\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199230 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db7591f0-3d80-418a-b4f3-850134650dea-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199253 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-plugins-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199291 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqwn\" (UniqueName: \"kubernetes.io/projected/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-kube-api-access-sgqwn\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199317 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-node-bootstrap-token\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199340 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8cd6278-992f-4d7b-81ce-9f444bac3730-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199362 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199381 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/80c7e7bf-f135-4728-b27b-0b23b8093b07-tmpfs\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199401 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e413497-e6a2-4e02-afb7-12be8584bb54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199422 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-socket-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199496 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad2193d1-bbed-4942-a2e5-40246e3c67de-metrics-tls\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199536 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199560 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7rwk\" (UniqueName: \"kubernetes.io/projected/f8ae1625-c133-48fc-b121-2490ae1e1a76-kube-api-access-x7rwk\") pod \"migrator-59844c95c7-8bmp9\" (UID: \"f8ae1625-c133-48fc-b121-2490ae1e1a76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199581 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgr5c\" (UniqueName: \"kubernetes.io/projected/80c7e7bf-f135-4728-b27b-0b23b8093b07-kube-api-access-cgr5c\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199603 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhjx\" (UniqueName: \"kubernetes.io/projected/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-kube-api-access-qfhjx\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199637 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199661 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2297\" (UniqueName: \"kubernetes.io/projected/630c4c0a-8622-4322-b217-49b25827730e-kube-api-access-x2297\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199684 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ba99723-c0fd-42ad-b926-93208dc9eba2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199708 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwhxr\" (UniqueName: \"kubernetes.io/projected/f8cd6278-992f-4d7b-81ce-9f444bac3730-kube-api-access-lwhxr\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199748 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984dfa93-96fc-401c-a7ad-61d60a9f77f2-config\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199775 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199810 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ead3bf87-613f-4624-9eaa-29fd11f02cfd-config\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199831 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgxr\" (UniqueName: \"kubernetes.io/projected/191860f0-d71a-44e8-b5ca-5eab39d390c0-kube-api-access-vmgxr\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199872 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6affa4-e748-447d-85d9-78c91b480940-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199895 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgp2x\" (UniqueName: \"kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199917 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvvch\" (UniqueName: \"kubernetes.io/projected/65baaf43-84aa-4f18-99ab-a7816e4600a4-kube-api-access-fvvch\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199939 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-srv-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199961 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.199981 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-images\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200006 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78ttt\" (UniqueName: \"kubernetes.io/projected/912c9374-00bd-47df-b8b5-5b2fc8e1675a-kube-api-access-78ttt\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200047 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba99723-c0fd-42ad-b926-93208dc9eba2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200115 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba99723-c0fd-42ad-b926-93208dc9eba2-config\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200150 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-bound-sa-token\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200188 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/912c9374-00bd-47df-b8b5-5b2fc8e1675a-metrics-tls\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200233 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-profile-collector-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200265 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-stats-auth\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.200289 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ead3bf87-613f-4624-9eaa-29fd11f02cfd-serving-cert\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.203113 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.203157 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crpk7\" (UniqueName: \"kubernetes.io/projected/48fd8d73-d00c-4512-9a1c-5669687a6f5c-kube-api-access-crpk7\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.203438 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/65baaf43-84aa-4f18-99ab-a7816e4600a4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.203460 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/630c4c0a-8622-4322-b217-49b25827730e-signing-key\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.205440 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.208381 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k4cjk"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.223968 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.281521 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.286437 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.299546 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.304730 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.304896 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwzpv\" (UniqueName: \"kubernetes.io/projected/ead3bf87-613f-4624-9eaa-29fd11f02cfd-kube-api-access-hwzpv\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.304949 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/984dfa93-96fc-401c-a7ad-61d60a9f77f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.304976 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e413497-e6a2-4e02-afb7-12be8584bb54-proxy-tls\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305001 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9h4p\" (UniqueName: \"kubernetes.io/projected/db7591f0-3d80-418a-b4f3-850134650dea-kube-api-access-h9h4p\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305023 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305058 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-webhook-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305081 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-certs\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305120 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk555\" (UniqueName: \"kubernetes.io/projected/d90c2cf7-520f-45f3-9c6d-e3970848f073-kube-api-access-qk555\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305152 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984dfa93-96fc-401c-a7ad-61d60a9f77f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305187 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwdkv\" (UniqueName: \"kubernetes.io/projected/5e413497-e6a2-4e02-afb7-12be8584bb54-kube-api-access-bwdkv\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305238 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c6affa4-e748-447d-85d9-78c91b480940-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305261 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ea85a4b-f44f-4500-94d8-99c116ab4e43-service-ca-bundle\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305288 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-metrics-tls\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305316 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db7591f0-3d80-418a-b4f3-850134650dea-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305340 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305364 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad2193d1-bbed-4942-a2e5-40246e3c67de-trusted-ca\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305386 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/191860f0-d71a-44e8-b5ca-5eab39d390c0-cert\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305411 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305436 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c6affa4-e748-447d-85d9-78c91b480940-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305461 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305486 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-csi-data-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305509 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-mountpoint-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305530 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-metrics-certs\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305555 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r56hl\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-kube-api-access-r56hl\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305591 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-proxy-tls\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305617 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpm6x\" (UniqueName: \"kubernetes.io/projected/6ea85a4b-f44f-4500-94d8-99c116ab4e43-kube-api-access-tpm6x\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305641 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkdk6\" (UniqueName: \"kubernetes.io/projected/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-kube-api-access-jkdk6\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305664 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjmp7\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305686 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305709 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305729 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-889vk\" (UniqueName: \"kubernetes.io/projected/8957b168-510a-4813-9036-900e09d378f5-kube-api-access-889vk\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305755 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-registration-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305777 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-default-certificate\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305800 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db7591f0-3d80-418a-b4f3-850134650dea-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305833 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-plugins-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305855 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-config-volume\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305880 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqwn\" (UniqueName: \"kubernetes.io/projected/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-kube-api-access-sgqwn\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305904 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-node-bootstrap-token\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305937 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8cd6278-992f-4d7b-81ce-9f444bac3730-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305955 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305971 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/80c7e7bf-f135-4728-b27b-0b23b8093b07-tmpfs\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.305986 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e413497-e6a2-4e02-afb7-12be8584bb54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306000 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-socket-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306018 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad2193d1-bbed-4942-a2e5-40246e3c67de-metrics-tls\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306035 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306050 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7rwk\" (UniqueName: \"kubernetes.io/projected/f8ae1625-c133-48fc-b121-2490ae1e1a76-kube-api-access-x7rwk\") pod \"migrator-59844c95c7-8bmp9\" (UID: \"f8ae1625-c133-48fc-b121-2490ae1e1a76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306064 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgr5c\" (UniqueName: \"kubernetes.io/projected/80c7e7bf-f135-4728-b27b-0b23b8093b07-kube-api-access-cgr5c\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306080 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhjx\" (UniqueName: \"kubernetes.io/projected/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-kube-api-access-qfhjx\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306094 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306109 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2297\" (UniqueName: \"kubernetes.io/projected/630c4c0a-8622-4322-b217-49b25827730e-kube-api-access-x2297\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306125 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ba99723-c0fd-42ad-b926-93208dc9eba2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306139 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwhxr\" (UniqueName: \"kubernetes.io/projected/f8cd6278-992f-4d7b-81ce-9f444bac3730-kube-api-access-lwhxr\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306154 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984dfa93-96fc-401c-a7ad-61d60a9f77f2-config\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306171 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306185 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ead3bf87-613f-4624-9eaa-29fd11f02cfd-config\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306200 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgxr\" (UniqueName: \"kubernetes.io/projected/191860f0-d71a-44e8-b5ca-5eab39d390c0-kube-api-access-vmgxr\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306246 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6affa4-e748-447d-85d9-78c91b480940-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306262 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgp2x\" (UniqueName: \"kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306278 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvvch\" (UniqueName: \"kubernetes.io/projected/65baaf43-84aa-4f18-99ab-a7816e4600a4-kube-api-access-fvvch\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306322 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-srv-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306339 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306355 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-images\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306373 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78ttt\" (UniqueName: \"kubernetes.io/projected/912c9374-00bd-47df-b8b5-5b2fc8e1675a-kube-api-access-78ttt\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306389 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba99723-c0fd-42ad-b926-93208dc9eba2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306406 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26zs\" (UniqueName: \"kubernetes.io/projected/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-kube-api-access-j26zs\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306429 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba99723-c0fd-42ad-b926-93208dc9eba2-config\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306447 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-bound-sa-token\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306467 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/912c9374-00bd-47df-b8b5-5b2fc8e1675a-metrics-tls\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306483 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-profile-collector-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306500 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-stats-auth\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306518 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ead3bf87-613f-4624-9eaa-29fd11f02cfd-serving-cert\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306550 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306566 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crpk7\" (UniqueName: \"kubernetes.io/projected/48fd8d73-d00c-4512-9a1c-5669687a6f5c-kube-api-access-crpk7\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306590 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/65baaf43-84aa-4f18-99ab-a7816e4600a4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306607 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/630c4c0a-8622-4322-b217-49b25827730e-signing-key\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306623 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8cd6278-992f-4d7b-81ce-9f444bac3730-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306653 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-srv-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.306672 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/630c4c0a-8622-4322-b217-49b25827730e-signing-cabundle\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.307606 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/630c4c0a-8622-4322-b217-49b25827730e-signing-cabundle\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.310589 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.311235 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.311676 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:12.811659849 +0000 UTC m=+139.678630236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.312132 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.312324 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8cd6278-992f-4d7b-81ce-9f444bac3730-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.312465 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/80c7e7bf-f135-4728-b27b-0b23b8093b07-tmpfs\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.312989 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e413497-e6a2-4e02-afb7-12be8584bb54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.313149 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-socket-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.313325 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.315666 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-images\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.316176 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-mountpoint-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.316316 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-csi-data-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.317170 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ea85a4b-f44f-4500-94d8-99c116ab4e43-service-ca-bundle\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.318180 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.318869 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984dfa93-96fc-401c-a7ad-61d60a9f77f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.319289 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-profile-collector-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.319758 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/912c9374-00bd-47df-b8b5-5b2fc8e1675a-metrics-tls\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.320641 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984dfa93-96fc-401c-a7ad-61d60a9f77f2-config\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.321247 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad2193d1-bbed-4942-a2e5-40246e3c67de-trusted-ca\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.321637 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c6affa4-e748-447d-85d9-78c91b480940-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322135 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db7591f0-3d80-418a-b4f3-850134650dea-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322753 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e413497-e6a2-4e02-afb7-12be8584bb54-proxy-tls\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322795 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6affa4-e748-447d-85d9-78c91b480940-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322877 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad2193d1-bbed-4942-a2e5-40246e3c67de-metrics-tls\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322889 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8cd6278-992f-4d7b-81ce-9f444bac3730-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.322971 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-plugins-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.323029 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/48fd8d73-d00c-4512-9a1c-5669687a6f5c-registration-dir\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.323392 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db7591f0-3d80-418a-b4f3-850134650dea-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.323607 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.323806 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d90c2cf7-520f-45f3-9c6d-e3970848f073-srv-cert\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.324593 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba99723-c0fd-42ad-b926-93208dc9eba2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.324851 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-srv-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.324987 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.332349 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-stats-auth\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.332692 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/630c4c0a-8622-4322-b217-49b25827730e-signing-key\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.332803 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.333099 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.333829 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-default-certificate\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.334095 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-node-bootstrap-token\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.334709 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/191860f0-d71a-44e8-b5ca-5eab39d390c0-cert\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.335265 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/65baaf43-84aa-4f18-99ab-a7816e4600a4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.335832 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ea85a4b-f44f-4500-94d8-99c116ab4e43-metrics-certs\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.340505 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-proxy-tls\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.340567 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8957b168-510a-4813-9036-900e09d378f5-certs\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.344849 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80c7e7bf-f135-4728-b27b-0b23b8093b07-webhook-cert\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.347901 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgp2x\" (UniqueName: \"kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x\") pod \"collect-profiles-29485965-g8clk\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.370336 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r6525"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.373918 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpm6x\" (UniqueName: \"kubernetes.io/projected/6ea85a4b-f44f-4500-94d8-99c116ab4e43-kube-api-access-tpm6x\") pod \"router-default-5444994796-hszft\" (UID: \"6ea85a4b-f44f-4500-94d8-99c116ab4e43\") " pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.376959 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhxvh"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.390191 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.400768 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78ttt\" (UniqueName: \"kubernetes.io/projected/912c9374-00bd-47df-b8b5-5b2fc8e1675a-kube-api-access-78ttt\") pod \"dns-operator-744455d44c-zjw59\" (UID: \"912c9374-00bd-47df-b8b5-5b2fc8e1675a\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.407597 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.407689 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-metrics-tls\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.407748 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-config-volume\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.407848 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j26zs\" (UniqueName: \"kubernetes.io/projected/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-kube-api-access-j26zs\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.408309 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:12.908195614 +0000 UTC m=+139.775165991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.416703 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvvch\" (UniqueName: \"kubernetes.io/projected/65baaf43-84aa-4f18-99ab-a7816e4600a4-kube-api-access-fvvch\") pod \"control-plane-machine-set-operator-78cbb6b69f-qfb94\" (UID: \"65baaf43-84aa-4f18-99ab-a7816e4600a4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.430056 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.436629 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkdk6\" (UniqueName: \"kubernetes.io/projected/28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7-kube-api-access-jkdk6\") pod \"machine-config-operator-74547568cd-dd4dm\" (UID: \"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.436892 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.467510 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-n7s8j"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.468999 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.473157 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.474608 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.494489 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r56hl\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-kube-api-access-r56hl\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.497620 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.508713 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.509186 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.009160739 +0000 UTC m=+139.876131126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.531457 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.532230 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/984dfa93-96fc-401c-a7ad-61d60a9f77f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v78lj\" (UID: \"984dfa93-96fc-401c-a7ad-61d60a9f77f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.548467 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.553313 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk555\" (UniqueName: \"kubernetes.io/projected/d90c2cf7-520f-45f3-9c6d-e3970848f073-kube-api-access-qk555\") pod \"catalog-operator-68c6474976-h2gr8\" (UID: \"d90c2cf7-520f-45f3-9c6d-e3970848f073\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.554436 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.592236 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwdkv\" (UniqueName: \"kubernetes.io/projected/5e413497-e6a2-4e02-afb7-12be8584bb54-kube-api-access-bwdkv\") pod \"machine-config-controller-84d6567774-c49bm\" (UID: \"5e413497-e6a2-4e02-afb7-12be8584bb54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.610122 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.610533 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.110515443 +0000 UTC m=+139.977485830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.612229 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c6affa4-e748-447d-85d9-78c91b480940-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8h4jf\" (UID: \"7c6affa4-e748-447d-85d9-78c91b480940\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.632729 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad2193d1-bbed-4942-a2e5-40246e3c67de-bound-sa-token\") pod \"ingress-operator-5b745b69d9-clf44\" (UID: \"ad2193d1-bbed-4942-a2e5-40246e3c67de\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.633720 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.650303 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.653788 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwhxr\" (UniqueName: \"kubernetes.io/projected/f8cd6278-992f-4d7b-81ce-9f444bac3730-kube-api-access-lwhxr\") pod \"kube-storage-version-migrator-operator-b67b599dd-krd2c\" (UID: \"f8cd6278-992f-4d7b-81ce-9f444bac3730\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.673885 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9h4p\" (UniqueName: \"kubernetes.io/projected/db7591f0-3d80-418a-b4f3-850134650dea-kube-api-access-h9h4p\") pod \"openshift-controller-manager-operator-756b6f6bc6-28lqh\" (UID: \"db7591f0-3d80-418a-b4f3-850134650dea\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.693724 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4hxzp"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.700011 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.700089 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.709999 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqzln"] Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.711444 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.711562 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.211540439 +0000 UTC m=+140.078510826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.711762 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.712130 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.212117404 +0000 UTC m=+140.079087791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.714644 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.722919 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.728117 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.731622 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhjx\" (UniqueName: \"kubernetes.io/projected/47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e-kube-api-access-qfhjx\") pod \"multus-admission-controller-857f4d67dd-6jqvj\" (UID: \"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.751634 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2297\" (UniqueName: \"kubernetes.io/projected/630c4c0a-8622-4322-b217-49b25827730e-kube-api-access-x2297\") pod \"service-ca-9c57cc56f-rzm66\" (UID: \"630c4c0a-8622-4322-b217-49b25827730e\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.790349 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.791343 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgxr\" (UniqueName: \"kubernetes.io/projected/191860f0-d71a-44e8-b5ca-5eab39d390c0-kube-api-access-vmgxr\") pod \"ingress-canary-flx89\" (UID: \"191860f0-d71a-44e8-b5ca-5eab39d390c0\") " pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.813180 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.813500 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.313472288 +0000 UTC m=+140.180442705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.813656 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.814033 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.314020232 +0000 UTC m=+140.180990629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.814536 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.833476 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-config-volume\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.833654 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-889vk\" (UniqueName: \"kubernetes.io/projected/8957b168-510a-4813-9036-900e09d378f5-kube-api-access-889vk\") pod \"machine-config-server-2f5n6\" (UID: \"8957b168-510a-4813-9036-900e09d378f5\") " pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.833673 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ead3bf87-613f-4624-9eaa-29fd11f02cfd-config\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.833818 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.834631 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba99723-c0fd-42ad-b926-93208dc9eba2-config\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.835362 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.835520 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.835964 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgr5c\" (UniqueName: \"kubernetes.io/projected/80c7e7bf-f135-4728-b27b-0b23b8093b07-kube-api-access-cgr5c\") pod \"packageserver-d55dfcdfc-nxzpb\" (UID: \"80c7e7bf-f135-4728-b27b-0b23b8093b07\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.837637 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwzpv\" (UniqueName: \"kubernetes.io/projected/ead3bf87-613f-4624-9eaa-29fd11f02cfd-kube-api-access-hwzpv\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.837832 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-metrics-tls\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.838175 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ba99723-c0fd-42ad-b926-93208dc9eba2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-p62dc\" (UID: \"4ba99723-c0fd-42ad-b926-93208dc9eba2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.838615 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ead3bf87-613f-4624-9eaa-29fd11f02cfd-serving-cert\") pod \"service-ca-operator-777779d784-gwvhs\" (UID: \"ead3bf87-613f-4624-9eaa-29fd11f02cfd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.838898 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crpk7\" (UniqueName: \"kubernetes.io/projected/48fd8d73-d00c-4512-9a1c-5669687a6f5c-kube-api-access-crpk7\") pod \"csi-hostpathplugin-cbtmt\" (UID: \"48fd8d73-d00c-4512-9a1c-5669687a6f5c\") " pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.838901 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqwn\" (UniqueName: \"kubernetes.io/projected/5fee33cd-f588-4cfe-a9a4-90fcb301bccb-kube-api-access-sgqwn\") pod \"olm-operator-6b444d44fb-p5wwg\" (UID: \"5fee33cd-f588-4cfe-a9a4-90fcb301bccb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.840318 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.840367 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.842098 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjmp7\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.845177 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7rwk\" (UniqueName: \"kubernetes.io/projected/f8ae1625-c133-48fc-b121-2490ae1e1a76-kube-api-access-x7rwk\") pod \"migrator-59844c95c7-8bmp9\" (UID: \"f8ae1625-c133-48fc-b121-2490ae1e1a76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.858709 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.866600 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2f5n6" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.876792 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-flx89" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.884662 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j26zs\" (UniqueName: \"kubernetes.io/projected/8b7e6a76-2267-4c5a-9137-3a50ca9a18ec-kube-api-access-j26zs\") pod \"dns-default-hgqbd\" (UID: \"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec\") " pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.894885 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.903700 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.914884 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:12 crc kubenswrapper[4899]: E0123 08:54:12.915318 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.415259283 +0000 UTC m=+140.282229670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.927290 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.960736 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" Jan 23 08:54:12 crc kubenswrapper[4899]: I0123 08:54:12.967703 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.006126 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.016271 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.016569 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.516554636 +0000 UTC m=+140.383525023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.049453 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ctjz9"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.107058 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.113775 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzm66"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.117772 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.117911 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.61788904 +0000 UTC m=+140.484859427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.118094 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.118476 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.618465364 +0000 UTC m=+140.485435751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: W0123 08:54:13.169405 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1258f97e_0f29_4f1d_8cf0_ceeb0a7c5ea0.slice/crio-fafbdb3a2bc3bbaae2b6592ce1f1fe7c56e3ad488089e09467f1f603d46168b3 WatchSource:0}: Error finding container fafbdb3a2bc3bbaae2b6592ce1f1fe7c56e3ad488089e09467f1f603d46168b3: Status 404 returned error can't find the container with id fafbdb3a2bc3bbaae2b6592ce1f1fe7c56e3ad488089e09467f1f603d46168b3 Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.190881 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-n7s8j" event={"ID":"27183ff6-88f1-42ef-9d0c-bbd4da6d2441","Type":"ContainerStarted","Data":"0601682473d861dc5f5ff8200d159829b967ea76fb14c581e7a425a5f0ee7c8e"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.192622 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" event={"ID":"9f959534-2339-44d2-82dc-82f6c8e8a893","Type":"ContainerStarted","Data":"ec62ef2ae22386649078b9bb77b74c2fd030dc51953c7563f0fe407bc19b4163"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.194140 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" event={"ID":"e0e33477-d346-481d-8e80-bb90445a0eb2","Type":"ContainerStarted","Data":"af10a894ce9e18d63f86bb86037295f67a6e0a136a3a8f2632cad345a5693d05"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.199826 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" event={"ID":"ec02f16b-18e4-4bf8-a921-2104a1f7ffff","Type":"ContainerStarted","Data":"a43f44cb2dad3619f5442b1c14842e51b8b3eb6608d47a72c248e0392469b953"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.218959 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.219356 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" event={"ID":"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2","Type":"ContainerStarted","Data":"ab691c53526e9e2fe3e2894b60adbc324142b34d62039517dc6f1bd195b5df4b"} Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.219784 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.719748807 +0000 UTC m=+140.586719194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.245270 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" event={"ID":"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0","Type":"ContainerStarted","Data":"fafbdb3a2bc3bbaae2b6592ce1f1fe7c56e3ad488089e09467f1f603d46168b3"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.247947 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" event={"ID":"433100b8-a3d5-4f8a-8e88-57026c742941","Type":"ContainerStarted","Data":"9c020486824a63e0144923c9201eaafc51009645294e9c447711f6fdfbf315f2"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.277101 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r6525" event={"ID":"0be4b82f-6b52-486a-8f68-88a6cd3a8317","Type":"ContainerStarted","Data":"59c9464667f93e8193d8acdab776cf15bf62d4b9e39c05d3b4217aed1421b0b2"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.282137 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" event={"ID":"f626d106-0870-45f8-ab43-31d92627d068","Type":"ContainerStarted","Data":"de9540be44f8fb59aeea048866294b450ac7cea496eb0d46bcf57ae09643559f"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.283689 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" event={"ID":"9ad27e9b-b533-485b-9ec1-ac004ed76611","Type":"ContainerStarted","Data":"fef247bd6788f6b200dc66e1af4ccb380fbf500725d137b52506ddc50bc6fc51"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.285463 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerStarted","Data":"7b395f2ab6583378458029cff2a5a8592d70fa9f6c13f1a628221c26d0710e08"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.290087 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6jqvj"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.292840 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" event={"ID":"3709c439-2199-443d-8e3d-250b85dfdf2a","Type":"ContainerStarted","Data":"692fc8e98d6856f527a7d76068ae71e3495645e84bf56183ed43e125aeba46f1"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.294763 4899 generic.go:334] "Generic (PLEG): container finished" podID="986b43bc-5ff7-486a-8539-f7568bc468f1" containerID="4fcb0a0f400ba8757ebc2d2a2eae76738969c8fe8a0b628addf58360e6f0feb7" exitCode=0 Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.294817 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" event={"ID":"986b43bc-5ff7-486a-8539-f7568bc468f1","Type":"ContainerDied","Data":"4fcb0a0f400ba8757ebc2d2a2eae76738969c8fe8a0b628addf58360e6f0feb7"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.294838 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" event={"ID":"986b43bc-5ff7-486a-8539-f7568bc468f1","Type":"ContainerStarted","Data":"904482cbd98fab91622c4d5578a08b065830bfcf3a16a3bcc3df963a8a40a160"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.296789 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" event={"ID":"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3","Type":"ContainerStarted","Data":"cc0e1cabdccc874190112d3fa613c762e962b521316cf64d64aef8ed9a2ede86"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.297827 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" event={"ID":"4a7de32e-498b-46e2-9de3-7313c34eab26","Type":"ContainerStarted","Data":"091646442a52f05f92b5d4df2a8d94935beafd3d180deec1bdd881b315ffab79"} Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.321368 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.322523 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.822508416 +0000 UTC m=+140.689478803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.422407 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.423053 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:13.92303895 +0000 UTC m=+140.790009337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.445427 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjw59"] Jan 23 08:54:13 crc kubenswrapper[4899]: W0123 08:54:13.486042 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod912c9374_00bd_47df_b8b5_5b2fc8e1675a.slice/crio-c23f0f9d36de58051a28a3906b212130008e4f44ba1af4a4f53916992f974606 WatchSource:0}: Error finding container c23f0f9d36de58051a28a3906b212130008e4f44ba1af4a4f53916992f974606: Status 404 returned error can't find the container with id c23f0f9d36de58051a28a3906b212130008e4f44ba1af4a4f53916992f974606 Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.523738 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.524016 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.024005715 +0000 UTC m=+140.890976102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.591059 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tbwm2"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.608645 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.618702 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb"] Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.629290 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.629909 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.129868461 +0000 UTC m=+140.996838848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.732857 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.733664 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.233650836 +0000 UTC m=+141.100621223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.834001 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.834423 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.334402515 +0000 UTC m=+141.201372902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.834828 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.835191 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.335181564 +0000 UTC m=+141.202151951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:13 crc kubenswrapper[4899]: I0123 08:54:13.936582 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:13 crc kubenswrapper[4899]: E0123 08:54:13.937006 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.43698636 +0000 UTC m=+141.303956747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.038858 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.039483 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.539471042 +0000 UTC m=+141.406441439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.140172 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.140469 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.640442127 +0000 UTC m=+141.507412514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.140741 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.141180 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.641168105 +0000 UTC m=+141.508138492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.248381 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.249047 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.749028551 +0000 UTC m=+141.615998938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.338990 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" event={"ID":"f8cd6278-992f-4d7b-81ce-9f444bac3730","Type":"ContainerStarted","Data":"81aa72281fdf84a29f215828d44e3b234cb0c60832ea7e6000e547e7c3a8e16f"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.349757 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" event={"ID":"926c5bb9-7f91-4ea9-860c-694ec9c5c4db","Type":"ContainerStarted","Data":"5d238aec6169a379acc2a217671a53af0209c3657cf0cf59d8e30413fc746aac"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.354154 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.354599 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.85458822 +0000 UTC m=+141.721558607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.355794 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" event={"ID":"912c9374-00bd-47df-b8b5-5b2fc8e1675a","Type":"ContainerStarted","Data":"c23f0f9d36de58051a28a3906b212130008e4f44ba1af4a4f53916992f974606"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.362747 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.389431 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" event={"ID":"9f959534-2339-44d2-82dc-82f6c8e8a893","Type":"ContainerStarted","Data":"24d64bb049869912d764aa8b174fd919f140511fb113c25967481188dba33401"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.390950 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.391525 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" event={"ID":"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e","Type":"ContainerStarted","Data":"a21d5a4644ee16cacbe62f79adcbb8ec80d82135db9a3c740e8d3483bee53810"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.393188 4899 generic.go:334] "Generic (PLEG): container finished" podID="3709c439-2199-443d-8e3d-250b85dfdf2a" containerID="6e7e4a3e63a24f8aeacc95e1011af56d92ad5e13194d30c31620aeb9f211b9ea" exitCode=0 Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.393238 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" event={"ID":"3709c439-2199-443d-8e3d-250b85dfdf2a","Type":"ContainerDied","Data":"6e7e4a3e63a24f8aeacc95e1011af56d92ad5e13194d30c31620aeb9f211b9ea"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.393645 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-flx89"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.396330 4899 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vfz69 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.396377 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.398147 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r6525" event={"ID":"0be4b82f-6b52-486a-8f68-88a6cd3a8317","Type":"ContainerStarted","Data":"3a27dc3c7c3663235b6728ad269708e69bd9873f0d2b3fc8479c640ea7c8d225"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.398863 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.400697 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:14 crc kubenswrapper[4899]: W0123 08:54:14.401800 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53 WatchSource:0}: Error finding container 6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53: Status 404 returned error can't find the container with id 6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53 Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.402266 4899 patch_prober.go:28] interesting pod/downloads-7954f5f757-r6525 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.402394 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r6525" podUID="0be4b82f-6b52-486a-8f68-88a6cd3a8317" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.402718 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" event={"ID":"ea762979-3e58-4ab7-bf2a-8b7f1a5ad8e3","Type":"ContainerStarted","Data":"30505e9adc03e1f1a6e7673e72bbd752806c1e62d14a6dd3eae64c13a6c6331b"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.406604 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm"] Jan 23 08:54:14 crc kubenswrapper[4899]: W0123 08:54:14.406676 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod984dfa93_96fc_401c_a7ad_61d60a9f77f2.slice/crio-5d3b913e4a4a60c00530c1b6df01ebd8c07eb8f5d3a284ff12843095c0ed7fb8 WatchSource:0}: Error finding container 5d3b913e4a4a60c00530c1b6df01ebd8c07eb8f5d3a284ff12843095c0ed7fb8: Status 404 returned error can't find the container with id 5d3b913e4a4a60c00530c1b6df01ebd8c07eb8f5d3a284ff12843095c0ed7fb8 Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.409312 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-n7s8j" event={"ID":"27183ff6-88f1-42ef-9d0c-bbd4da6d2441","Type":"ContainerStarted","Data":"9313d7d1f7e7e13582c1510788d21064b8952e09466a83c6d906cf3b834f9b8e"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.411926 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" event={"ID":"b289e2ee-7e64-440d-9ab2-a20d3e6a5aa2","Type":"ContainerStarted","Data":"b39208353e6fa9078d546bb8a82c7a5ed3844ae21d82e8bdbe858cec1230807e"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.414139 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" event={"ID":"9ad27e9b-b533-485b-9ec1-ac004ed76611","Type":"ContainerStarted","Data":"d13c0633e20c71139716d2e7a024e44eee2cb303685497d1a7e1888bfd52e844"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.434170 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerStarted","Data":"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.435046 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.436958 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2f5n6" event={"ID":"8957b168-510a-4813-9036-900e09d378f5","Type":"ContainerStarted","Data":"aa2a5f7ae5321076532b97f786ac9832f4cd41cbe84f8fcd9edfcd3fa3094094"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.441835 4899 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wcqlx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.441880 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.442811 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hszft" event={"ID":"6ea85a4b-f44f-4500-94d8-99c116ab4e43","Type":"ContainerStarted","Data":"c307af4a03f7c13a44da65e3acca280024c46c1ab0e63989276c69e6ff182905"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.449122 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" event={"ID":"80c7e7bf-f135-4728-b27b-0b23b8093b07","Type":"ContainerStarted","Data":"db7e87d0d178800ed6dc3a4adee7224a9ad146bcfe8b933d164d3785cfe514d5"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.456914 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.458577 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:14.958553179 +0000 UTC m=+141.825523616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.481423 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" event={"ID":"ec02f16b-18e4-4bf8-a921-2104a1f7ffff","Type":"ContainerStarted","Data":"4bb5e34c14a84ecc540f335416141778e94bf71af72045e4c6e60cacd4d91b97"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.483106 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" event={"ID":"630c4c0a-8622-4322-b217-49b25827730e","Type":"ContainerStarted","Data":"101afe8b378b60bff9a859685bab0ece9d4eadab804e77daee560a3b23f61c34"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.484594 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" event={"ID":"f626d106-0870-45f8-ab43-31d92627d068","Type":"ContainerStarted","Data":"f75e75cab3084433dbbbe5bbadc9aa61ee370493a0e91e7fd9d2d5c93d44a783"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.487081 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" event={"ID":"4a7de32e-498b-46e2-9de3-7313c34eab26","Type":"ContainerStarted","Data":"dc214fcceaaaa992dbef30d5c4683ec0868bf9dc1fc4cf7fcb7e99bab37f4e5b"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.487723 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.495600 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" event={"ID":"433100b8-a3d5-4f8a-8e88-57026c742941","Type":"ContainerStarted","Data":"4e6a2572ac0bc72f3a1cc5e09670fe70fc4136a5662a428b8e3645596e68a385"} Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.530973 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.554732 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.558742 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.561128 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.061111893 +0000 UTC m=+141.928082280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.662598 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.162551579 +0000 UTC m=+142.029521966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.662840 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.663454 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.663768 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.163760359 +0000 UTC m=+142.030730746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.714591 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" podStartSLOduration=117.71457503 podStartE2EDuration="1m57.71457503s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.708581591 +0000 UTC m=+141.575551978" watchObservedRunningTime="2026-01-23 08:54:14.71457503 +0000 UTC m=+141.581545417" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.748034 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-555np" podStartSLOduration=117.74801611 podStartE2EDuration="1m57.74801611s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.747268041 +0000 UTC m=+141.614238428" watchObservedRunningTime="2026-01-23 08:54:14.74801611 +0000 UTC m=+141.614986497" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.765665 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.766079 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.266062537 +0000 UTC m=+142.133032924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.799824 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podStartSLOduration=116.799803584 podStartE2EDuration="1m56.799803584s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.798654716 +0000 UTC m=+141.665625103" watchObservedRunningTime="2026-01-23 08:54:14.799803584 +0000 UTC m=+141.666773971" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.833546 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bwh2k" podStartSLOduration=117.833530361 podStartE2EDuration="1m57.833530361s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.832031814 +0000 UTC m=+141.699002201" watchObservedRunningTime="2026-01-23 08:54:14.833530361 +0000 UTC m=+141.700500748" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.867221 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.867654 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.367639447 +0000 UTC m=+142.234609834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.894307 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podStartSLOduration=116.894276488 podStartE2EDuration="1m56.894276488s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.868842987 +0000 UTC m=+141.735813374" watchObservedRunningTime="2026-01-23 08:54:14.894276488 +0000 UTC m=+141.761246875" Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.897301 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.899268 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk"] Jan 23 08:54:14 crc kubenswrapper[4899]: W0123 08:54:14.908515 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28d7d3f8_0ab5_4ed0_ac86_eb67336cd4c7.slice/crio-347eb362821041d368df4f5920c4023d4fa128112ecbf263d0eef0b3c5a6382a WatchSource:0}: Error finding container 347eb362821041d368df4f5920c4023d4fa128112ecbf263d0eef0b3c5a6382a: Status 404 returned error can't find the container with id 347eb362821041d368df4f5920c4023d4fa128112ecbf263d0eef0b3c5a6382a Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.919601 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg"] Jan 23 08:54:14 crc kubenswrapper[4899]: W0123 08:54:14.921349 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5eca0f4_a14b_436b_8d5e_045d7214deee.slice/crio-9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604 WatchSource:0}: Error finding container 9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604: Status 404 returned error can't find the container with id 9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604 Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.957281 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.957338 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-clf44"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.972730 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8"] Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.973892 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:14 crc kubenswrapper[4899]: E0123 08:54:14.975184 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.475155364 +0000 UTC m=+142.342125761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:14 crc kubenswrapper[4899]: I0123 08:54:14.997745 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h5tlv"] Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.004479 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh"] Jan 23 08:54:15 crc kubenswrapper[4899]: W0123 08:54:15.006591 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8ae1625_c133_48fc_b121_2490ae1e1a76.slice/crio-cc7e74c5395a140b22addb0f9f222c7bc915c9bdefc9bfa95087dddac166ff89 WatchSource:0}: Error finding container cc7e74c5395a140b22addb0f9f222c7bc915c9bdefc9bfa95087dddac166ff89: Status 404 returned error can't find the container with id cc7e74c5395a140b22addb0f9f222c7bc915c9bdefc9bfa95087dddac166ff89 Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.011580 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hgqbd"] Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.015148 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc"] Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.026930 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9"] Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.030906 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cbtmt"] Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.037868 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-n7s8j" podStartSLOduration=118.037836349 podStartE2EDuration="1m58.037836349s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:14.985681365 +0000 UTC m=+141.852651752" watchObservedRunningTime="2026-01-23 08:54:15.037836349 +0000 UTC m=+141.904806736" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.042355 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" podStartSLOduration=118.042343311 podStartE2EDuration="1m58.042343311s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:15.031450101 +0000 UTC m=+141.898420488" watchObservedRunningTime="2026-01-23 08:54:15.042343311 +0000 UTC m=+141.909313698" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.075897 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-62gn5" podStartSLOduration=118.075878723 podStartE2EDuration="1m58.075878723s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:15.075429782 +0000 UTC m=+141.942400159" watchObservedRunningTime="2026-01-23 08:54:15.075878723 +0000 UTC m=+141.942849140" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.075989 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.076353 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.576305434 +0000 UTC m=+142.443275821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.077591 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-r6525" podStartSLOduration=118.077584525 podStartE2EDuration="1m58.077584525s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:15.062528852 +0000 UTC m=+141.929499239" watchObservedRunningTime="2026-01-23 08:54:15.077584525 +0000 UTC m=+141.944554912" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.177064 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.177477 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.677457973 +0000 UTC m=+142.544428360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.278407 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.278789 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.778770706 +0000 UTC m=+142.645741093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: W0123 08:54:15.307593 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b7e6a76_2267_4c5a_9137_3a50ca9a18ec.slice/crio-c5497b64b5075c1c17e3a7d944dee1bc4de4246232b9c6e3dd97408492187cb1 WatchSource:0}: Error finding container c5497b64b5075c1c17e3a7d944dee1bc4de4246232b9c6e3dd97408492187cb1: Status 404 returned error can't find the container with id c5497b64b5075c1c17e3a7d944dee1bc4de4246232b9c6e3dd97408492187cb1 Jan 23 08:54:15 crc kubenswrapper[4899]: W0123 08:54:15.314690 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd90c2cf7_520f_45f3_9c6d_e3970848f073.slice/crio-2701f89da95553088d6e67c2a3cf33a85c822fcb88f9a1acd128ea220437bbe9 WatchSource:0}: Error finding container 2701f89da95553088d6e67c2a3cf33a85c822fcb88f9a1acd128ea220437bbe9: Status 404 returned error can't find the container with id 2701f89da95553088d6e67c2a3cf33a85c822fcb88f9a1acd128ea220437bbe9 Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.379849 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.379993 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.879975467 +0000 UTC m=+142.746945854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.380192 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.380440 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.880433348 +0000 UTC m=+142.747403735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: W0123 08:54:15.427364 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb7591f0_3d80_418a_b4f3_850134650dea.slice/crio-00ad7ff82cbe896b22b33bfa8f25d5573e7ba53cd33d5ba43e4bbb77f9b92616 WatchSource:0}: Error finding container 00ad7ff82cbe896b22b33bfa8f25d5573e7ba53cd33d5ba43e4bbb77f9b92616: Status 404 returned error can't find the container with id 00ad7ff82cbe896b22b33bfa8f25d5573e7ba53cd33d5ba43e4bbb77f9b92616 Jan 23 08:54:15 crc kubenswrapper[4899]: W0123 08:54:15.430066 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48fd8d73_d00c_4512_9a1c_5669687a6f5c.slice/crio-ddf9cb5e5d3f59abd900887467c4c47d07f570c960bf68ceabaf7de780f88000 WatchSource:0}: Error finding container ddf9cb5e5d3f59abd900887467c4c47d07f570c960bf68ceabaf7de780f88000: Status 404 returned error can't find the container with id ddf9cb5e5d3f59abd900887467c4c47d07f570c960bf68ceabaf7de780f88000 Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.482635 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.482851 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.982824278 +0000 UTC m=+142.849794675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.483239 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.483571 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:15.983556966 +0000 UTC m=+142.850527353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.488435 4899 patch_prober.go:28] interesting pod/console-operator-58897d9998-4hxzp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded" start-of-body= Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.488493 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" podUID="4a7de32e-498b-46e2-9de3-7313c34eab26" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.521674 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" event={"ID":"db7591f0-3d80-418a-b4f3-850134650dea","Type":"ContainerStarted","Data":"00ad7ff82cbe896b22b33bfa8f25d5573e7ba53cd33d5ba43e4bbb77f9b92616"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.522332 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" event={"ID":"2376228d-4a51-459a-ad64-d152e53bf109","Type":"ContainerStarted","Data":"1d8e7cb2df7401d88b6b37df4ca94b64410f0de8d0ad46f359e35bdae5c98f57"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.523522 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" event={"ID":"5fee33cd-f588-4cfe-a9a4-90fcb301bccb","Type":"ContainerStarted","Data":"8d3afb91c0c692843d260f31f73d971f6802178ada7bd9ac8c4351fdae57a762"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.524606 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" event={"ID":"984dfa93-96fc-401c-a7ad-61d60a9f77f2","Type":"ContainerStarted","Data":"5d3b913e4a4a60c00530c1b6df01ebd8c07eb8f5d3a284ff12843095c0ed7fb8"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.525990 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" event={"ID":"7c6affa4-e748-447d-85d9-78c91b480940","Type":"ContainerStarted","Data":"b56f82f6a9d6d46541be059d0bd3d87a4591f0a6c7ac26b499c7cb3bfea9a433"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.528358 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" event={"ID":"d90c2cf7-520f-45f3-9c6d-e3970848f073","Type":"ContainerStarted","Data":"2701f89da95553088d6e67c2a3cf33a85c822fcb88f9a1acd128ea220437bbe9"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.530766 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" event={"ID":"8a08d77e-7f9c-4c41-8d85-95b64e461ea5","Type":"ContainerStarted","Data":"6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.538754 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" event={"ID":"5e413497-e6a2-4e02-afb7-12be8584bb54","Type":"ContainerStarted","Data":"7fdb5770b4357252393a6d74ec393119c0ebbad493d6dee2fc8157fa545751db"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.539761 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" event={"ID":"4ba99723-c0fd-42ad-b926-93208dc9eba2","Type":"ContainerStarted","Data":"17d84f6d5cd5e7c2a4eec290609a6bb78efd69e2ca60cf09d9913090bfe9652a"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.541503 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" event={"ID":"48fd8d73-d00c-4512-9a1c-5669687a6f5c","Type":"ContainerStarted","Data":"ddf9cb5e5d3f59abd900887467c4c47d07f570c960bf68ceabaf7de780f88000"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.544499 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" event={"ID":"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7","Type":"ContainerStarted","Data":"347eb362821041d368df4f5920c4023d4fa128112ecbf263d0eef0b3c5a6382a"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.545728 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" event={"ID":"f8ae1625-c133-48fc-b121-2490ae1e1a76","Type":"ContainerStarted","Data":"cc7e74c5395a140b22addb0f9f222c7bc915c9bdefc9bfa95087dddac166ff89"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.549144 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" event={"ID":"e5eca0f4-a14b-436b-8d5e-045d7214deee","Type":"ContainerStarted","Data":"9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.550128 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgqbd" event={"ID":"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec","Type":"ContainerStarted","Data":"c5497b64b5075c1c17e3a7d944dee1bc4de4246232b9c6e3dd97408492187cb1"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.551644 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" event={"ID":"65baaf43-84aa-4f18-99ab-a7816e4600a4","Type":"ContainerStarted","Data":"1c2a79edc2c8f26f030fe6a327f6a1d33424d7b4b1ffdf4a88821b5f2cbfdca1"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.552684 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-flx89" event={"ID":"191860f0-d71a-44e8-b5ca-5eab39d390c0","Type":"ContainerStarted","Data":"ee9730a2da8e071acea440e2ea708faba9c79c20d3cd96ac53f2b928dcbaf0a8"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.553725 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" event={"ID":"ad2193d1-bbed-4942-a2e5-40246e3c67de","Type":"ContainerStarted","Data":"79b588f3d20097994c02273fceb15df949910a015f1d057c576e20bc801b901a"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.555393 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" event={"ID":"ead3bf87-613f-4624-9eaa-29fd11f02cfd","Type":"ContainerStarted","Data":"4dd8a3e986b716745a8775c40d31e884ca20e8790e39455cf725778cba82022d"} Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.555604 4899 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wcqlx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.555643 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.556753 4899 patch_prober.go:28] interesting pod/downloads-7954f5f757-r6525 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.556804 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r6525" podUID="0be4b82f-6b52-486a-8f68-88a6cd3a8317" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.557024 4899 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vfz69 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.557049 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.584409 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.584566 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.084521841 +0000 UTC m=+142.951492238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.586364 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.586897 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.08687875 +0000 UTC m=+142.953849137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.687973 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.688188 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.188160632 +0000 UTC m=+143.055131019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.688320 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.688693 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.188679485 +0000 UTC m=+143.055649872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.789728 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.789897 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.289875945 +0000 UTC m=+143.156846342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.790083 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.790446 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.290436679 +0000 UTC m=+143.157407086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.891389 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.891554 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.391538527 +0000 UTC m=+143.258508914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.891754 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.892061 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.392055009 +0000 UTC m=+143.259025396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.924636 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4hxzp" Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.928895 4899 csr.go:261] certificate signing request csr-4jr4m is approved, waiting to be issued Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.937540 4899 csr.go:257] certificate signing request csr-4jr4m is issued Jan 23 08:54:15 crc kubenswrapper[4899]: I0123 08:54:15.992554 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:15 crc kubenswrapper[4899]: E0123 08:54:15.992951 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.492934722 +0000 UTC m=+143.359905109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.093279 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.093580 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.593569538 +0000 UTC m=+143.460539925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.196802 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.196992 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.696956773 +0000 UTC m=+143.563927160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.197702 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.198401 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.698380269 +0000 UTC m=+143.565350676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.300894 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.301408 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.801389934 +0000 UTC m=+143.668360321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.402883 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.403522 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:16.903505927 +0000 UTC m=+143.770476314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.506515 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.507007 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.006990655 +0000 UTC m=+143.873961042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.587660 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" event={"ID":"9ad27e9b-b533-485b-9ec1-ac004ed76611","Type":"ContainerStarted","Data":"6df7e0d2cefb3c0b8e1d1bb426ca9c817d67a45b9dfabe34335150b80b2c280c"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.609192 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.609859 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.109832516 +0000 UTC m=+143.976802903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.634123 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" event={"ID":"f8cd6278-992f-4d7b-81ce-9f444bac3730","Type":"ContainerStarted","Data":"bcff889269261f9c6e88259737fc80e96a2de816b814894cefc9d741adc6ea09"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.651019 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" event={"ID":"e5eca0f4-a14b-436b-8d5e-045d7214deee","Type":"ContainerStarted","Data":"62762c4c938b183f7cad91a1b3b8c43e5d30af97182033f5f3a5c4b9e1dabedc"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.668414 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" event={"ID":"ec02f16b-18e4-4bf8-a921-2104a1f7ffff","Type":"ContainerStarted","Data":"4fb1d4fe755e8b6436ff57297fcc1bba668447cfbc5b21013c766a5f73cd52f8"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.677743 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-flx89" event={"ID":"191860f0-d71a-44e8-b5ca-5eab39d390c0","Type":"ContainerStarted","Data":"91ceaa83d567da768819d657d6a4e7645de0f2765baeb89c5b4f9078cbaaa1b0"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.689642 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" podStartSLOduration=119.689627385 podStartE2EDuration="1m59.689627385s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.688724493 +0000 UTC m=+143.555694880" watchObservedRunningTime="2026-01-23 08:54:16.689627385 +0000 UTC m=+143.556597772" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.699574 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" event={"ID":"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e","Type":"ContainerStarted","Data":"c7611d326ffe1471533ee05adcabb6a25251656f0225e6cb06d1340e429bfda9"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.710001 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2f5n6" event={"ID":"8957b168-510a-4813-9036-900e09d378f5","Type":"ContainerStarted","Data":"9edd6f28e8b025afbc3a830d0114100dcd12b3145809cf037ff3d5fc9b7a4a85"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.711734 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.712339 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.212289467 +0000 UTC m=+144.079259874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.719005 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lnnll" podStartSLOduration=119.718985694 podStartE2EDuration="1m59.718985694s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.717964758 +0000 UTC m=+143.584935145" watchObservedRunningTime="2026-01-23 08:54:16.718985694 +0000 UTC m=+143.585956081" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.737894 4899 generic.go:334] "Generic (PLEG): container finished" podID="e0e33477-d346-481d-8e80-bb90445a0eb2" containerID="fca1f2885f68fd480ace68e761f377d4d6518c056e68daaff4b9c338a0f78ef6" exitCode=0 Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.738002 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" event={"ID":"e0e33477-d346-481d-8e80-bb90445a0eb2","Type":"ContainerDied","Data":"fca1f2885f68fd480ace68e761f377d4d6518c056e68daaff4b9c338a0f78ef6"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.746332 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" event={"ID":"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0","Type":"ContainerStarted","Data":"a46826f41d8c9d5ae450062b940e8aa27584cc5bf990d48c3ce55b7f7a6e284f"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.747848 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.750537 4899 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ctjz9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.750592 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.753822 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" event={"ID":"986b43bc-5ff7-486a-8539-f7568bc468f1","Type":"ContainerStarted","Data":"b73e448c2d993d7b52a93c8b4fbbbb7f32bb7b257574e6b7aa5b119d1e0ece82"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.759088 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" event={"ID":"8a08d77e-7f9c-4c41-8d85-95b64e461ea5","Type":"ContainerStarted","Data":"05a5fc672ac3db17e00405398b1a20edb90d73fcc4c9a87b9d33766b6250d960"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.759806 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.761946 4899 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2xtw6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.762001 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.769807 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hszft" event={"ID":"6ea85a4b-f44f-4500-94d8-99c116ab4e43","Type":"ContainerStarted","Data":"c5e9e8667e770eb2aa3ed1323567bcd25dd78cf469073349ca3a442a1e4c81c2"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.771771 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-flx89" podStartSLOduration=7.771754913 podStartE2EDuration="7.771754913s" podCreationTimestamp="2026-01-23 08:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.742896067 +0000 UTC m=+143.609866454" watchObservedRunningTime="2026-01-23 08:54:16.771754913 +0000 UTC m=+143.638725300" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.772273 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2f5n6" podStartSLOduration=7.772267225 podStartE2EDuration="7.772267225s" podCreationTimestamp="2026-01-23 08:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.770713737 +0000 UTC m=+143.637684134" watchObservedRunningTime="2026-01-23 08:54:16.772267225 +0000 UTC m=+143.639237612" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.781556 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" event={"ID":"65baaf43-84aa-4f18-99ab-a7816e4600a4","Type":"ContainerStarted","Data":"024dd824662e3cf35d8f50ab0834d022ef99d073234acb53869f9e138c4fd79b"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.796118 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" event={"ID":"630c4c0a-8622-4322-b217-49b25827730e","Type":"ContainerStarted","Data":"f4cec662fba219ecf4d644bd45d495f93406c8148d524607869899ac28d47de8"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.803090 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" event={"ID":"984dfa93-96fc-401c-a7ad-61d60a9f77f2","Type":"ContainerStarted","Data":"174fa53c6f7fb3b3e824e9bb804dc7877c8f447c99e09f67913c753a9588d66a"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.815080 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" event={"ID":"7c6affa4-e748-447d-85d9-78c91b480940","Type":"ContainerStarted","Data":"5650548a9052abf132f217badfaf093c8ee7ae6683ae4d4e3f0c4cfaeb533a36"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.816600 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.817702 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.317681952 +0000 UTC m=+144.184652349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.833646 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" event={"ID":"3709c439-2199-443d-8e3d-250b85dfdf2a","Type":"ContainerStarted","Data":"0a0321475e13f0acc1540cad7b151bea160a64fcf6cfe98600e09184302540cb"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.834205 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.840158 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" event={"ID":"912c9374-00bd-47df-b8b5-5b2fc8e1675a","Type":"ContainerStarted","Data":"abb13af2bc42bc5297fb27ab440ffd74fc111886a67df8ac8bd1812b86768d20"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.853004 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgqbd" event={"ID":"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec","Type":"ContainerStarted","Data":"b84fa9d279e225655253cd34a3efae1f00ddda988e942f6bb73c671839d8e32a"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.856250 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" event={"ID":"80c7e7bf-f135-4728-b27b-0b23b8093b07","Type":"ContainerStarted","Data":"c0a58cbd9516c120c6517cb9ac6b19fc4f5a08a45e781b0fce5f812c52dcc273"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.856725 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.860609 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" event={"ID":"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7","Type":"ContainerStarted","Data":"3e63aa92f6370c24c8aa247ab76582568bb9c4d5b4116e0a1f57c2f783433383"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.863607 4899 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nxzpb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.863662 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" podUID="80c7e7bf-f135-4728-b27b-0b23b8093b07" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.887442 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" event={"ID":"926c5bb9-7f91-4ea9-860c-694ec9c5c4db","Type":"ContainerStarted","Data":"51a9cdb2b00a5dd8b923443eaee3b019b57ad9fe2a07a72de902b5060568429d"} Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.888336 4899 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wcqlx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.888380 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.917565 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:16 crc kubenswrapper[4899]: E0123 08:54:16.919080 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.419065857 +0000 UTC m=+144.286036244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.943994 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-23 08:49:15 +0000 UTC, rotation deadline is 2026-11-13 09:48:54.456401891 +0000 UTC Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.944034 4899 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7056h54m37.512370285s for next certificate rotation Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.955343 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" podStartSLOduration=119.955322836 podStartE2EDuration="1m59.955322836s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.874581394 +0000 UTC m=+143.741551801" watchObservedRunningTime="2026-01-23 08:54:16.955322836 +0000 UTC m=+143.822293223" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.956419 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" podStartSLOduration=118.956411764 podStartE2EDuration="1m58.956411764s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.950960588 +0000 UTC m=+143.817930975" watchObservedRunningTime="2026-01-23 08:54:16.956411764 +0000 UTC m=+143.823382151" Jan 23 08:54:16 crc kubenswrapper[4899]: I0123 08:54:16.978039 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qfb94" podStartSLOduration=118.9780205 podStartE2EDuration="1m58.9780205s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:16.977606049 +0000 UTC m=+143.844576436" watchObservedRunningTime="2026-01-23 08:54:16.9780205 +0000 UTC m=+143.844990887" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.024224 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.025007 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.524993535 +0000 UTC m=+144.391963922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.073290 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rzm66" podStartSLOduration=119.073270713 podStartE2EDuration="1m59.073270713s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.04251176 +0000 UTC m=+143.909482147" watchObservedRunningTime="2026-01-23 08:54:17.073270713 +0000 UTC m=+143.940241100" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.097641 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-hszft" podStartSLOduration=119.097624217 podStartE2EDuration="1m59.097624217s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.073095508 +0000 UTC m=+143.940065895" watchObservedRunningTime="2026-01-23 08:54:17.097624217 +0000 UTC m=+143.964594604" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.098275 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8h4jf" podStartSLOduration=119.098270023 podStartE2EDuration="1m59.098270023s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.096079768 +0000 UTC m=+143.963050155" watchObservedRunningTime="2026-01-23 08:54:17.098270023 +0000 UTC m=+143.965240410" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.126029 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.126889 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.626856322 +0000 UTC m=+144.493826729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.126990 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.127562 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.627546339 +0000 UTC m=+144.494516726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.202481 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" podStartSLOduration=119.202460357 podStartE2EDuration="1m59.202460357s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.152042927 +0000 UTC m=+144.019013324" watchObservedRunningTime="2026-01-23 08:54:17.202460357 +0000 UTC m=+144.069430744" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.231962 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.232654 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.732636006 +0000 UTC m=+144.599606393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.253613 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.287869 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v78lj" podStartSLOduration=119.287850936 podStartE2EDuration="1m59.287850936s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.202415856 +0000 UTC m=+144.069386243" watchObservedRunningTime="2026-01-23 08:54:17.287850936 +0000 UTC m=+144.154821323" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.288776 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" podStartSLOduration=119.288770159 podStartE2EDuration="1m59.288770159s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.286665776 +0000 UTC m=+144.153636183" watchObservedRunningTime="2026-01-23 08:54:17.288770159 +0000 UTC m=+144.155740546" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.337224 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.337639 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.837626131 +0000 UTC m=+144.704596518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.361660 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" podStartSLOduration=120.361638296 podStartE2EDuration="2m0.361638296s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.359068952 +0000 UTC m=+144.226039349" watchObservedRunningTime="2026-01-23 08:54:17.361638296 +0000 UTC m=+144.228608683" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.448437 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.448848 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:17.948833539 +0000 UTC m=+144.815803926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.550264 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.550682 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.050667806 +0000 UTC m=+144.917638203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.637603 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.654016 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.654123 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:17 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:17 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:17 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.654177 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.654694 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.154674316 +0000 UTC m=+145.021644703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.755555 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.755840 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.255829035 +0000 UTC m=+145.122799422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.856542 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.856771 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.356744349 +0000 UTC m=+145.223714736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.856879 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.857289 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.357278632 +0000 UTC m=+145.224249019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.914570 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" event={"ID":"e0e33477-d346-481d-8e80-bb90445a0eb2","Type":"ContainerStarted","Data":"2438ea711d88d014470dc5e7daf74fd4e4a4328b0617e03099bc70e0cc814943"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.916575 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-p62dc" event={"ID":"4ba99723-c0fd-42ad-b926-93208dc9eba2","Type":"ContainerStarted","Data":"addb5954e56a3fb7db58ef169de5d05a9645bb035776812694cbbcc149990b57"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.919497 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" event={"ID":"2376228d-4a51-459a-ad64-d152e53bf109","Type":"ContainerStarted","Data":"ecd3ea1c8d30a5da7661f094014396169391a2a06c4963d17d4bdac4ba9a19c9"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.921348 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgqbd" event={"ID":"8b7e6a76-2267-4c5a-9137-3a50ca9a18ec","Type":"ContainerStarted","Data":"ba823dc0e28455786ef99c1ae6cdd49c2ec036164297357741dbcde4bcbccff3"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.921475 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.923411 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" event={"ID":"912c9374-00bd-47df-b8b5-5b2fc8e1675a","Type":"ContainerStarted","Data":"8064b56fe0031c0f79aca7a6882adcb42d552f3c831e2bc3b30f0191d3207a4a"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.928899 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" event={"ID":"47b3ebc2-2937-4c6e-b9b4-9374bfa9d54e","Type":"ContainerStarted","Data":"c0c71b36262d2c9bbe45c8566f5bdcd33eff8158e5e9738d43efd32a414b57c3"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.946394 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" event={"ID":"986b43bc-5ff7-486a-8539-f7568bc468f1","Type":"ContainerStarted","Data":"adaec1a13f5168c7f66cdc58b78d085aed7527a1e6b6e744464873d1b238e99e"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.958019 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.958326 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.458294948 +0000 UTC m=+145.325265345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.958665 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:17 crc kubenswrapper[4899]: E0123 08:54:17.959867 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.459851797 +0000 UTC m=+145.326822184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.961920 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" event={"ID":"28d7d3f8-0ab5-4ed0-ac86-eb67336cd4c7","Type":"ContainerStarted","Data":"904b2a1e9146c66e1403f89d1724aeaf41d3c72f849968800bf29c770a32d503"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.980979 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" podStartSLOduration=119.98096377 podStartE2EDuration="1m59.98096377s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:17.978650463 +0000 UTC m=+144.845620860" watchObservedRunningTime="2026-01-23 08:54:17.98096377 +0000 UTC m=+144.847934157" Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.984895 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" event={"ID":"926c5bb9-7f91-4ea9-860c-694ec9c5c4db","Type":"ContainerStarted","Data":"53b433d5278904449c8639053abedc7cc3921b1a8a25cf54f2a8f1f118706924"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.999372 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" event={"ID":"f8ae1625-c133-48fc-b121-2490ae1e1a76","Type":"ContainerStarted","Data":"ac19440ca1f6829f2cd93885f34885545b0691692b01a7fef4e7b24d0c1c5d3a"} Jan 23 08:54:17 crc kubenswrapper[4899]: I0123 08:54:17.999715 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" event={"ID":"f8ae1625-c133-48fc-b121-2490ae1e1a76","Type":"ContainerStarted","Data":"f2be6268c75b59442e08f741f1798672f7ca62b8b2db887531825c632fb9bf86"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.026481 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" event={"ID":"5e413497-e6a2-4e02-afb7-12be8584bb54","Type":"ContainerStarted","Data":"2b8e2afcaf814cfc3b115a0a3cc1ef8a36ab126693a651bcac1b421dac71c8d1"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.026530 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" event={"ID":"5e413497-e6a2-4e02-afb7-12be8584bb54","Type":"ContainerStarted","Data":"83b832af668022b224c73037c63bfbb566bebecf5dcea01cc667167ab2bbc52d"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.039256 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" event={"ID":"db7591f0-3d80-418a-b4f3-850134650dea","Type":"ContainerStarted","Data":"667093ab2fdf18a10102a4ac0aaa49f48bcd57f011048c88022d1c4f99a47d78"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.055621 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" event={"ID":"d90c2cf7-520f-45f3-9c6d-e3970848f073","Type":"ContainerStarted","Data":"a5a3c194c427a570ce957123511f0ac921b35505e5ba1158c3cb8346481250eb"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.056479 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.059943 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.061331 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.561316674 +0000 UTC m=+145.428287061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.076856 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" event={"ID":"ad2193d1-bbed-4942-a2e5-40246e3c67de","Type":"ContainerStarted","Data":"45bed43a18fa379b3ca31f4759eba7038d8899d1a130f4635c610689a97749ce"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.076926 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" event={"ID":"ad2193d1-bbed-4942-a2e5-40246e3c67de","Type":"ContainerStarted","Data":"ea2649ce90500703b7b0f95929a9a9b23431eb82273b3bba55c24a204d2f7b4c"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.089401 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" event={"ID":"ead3bf87-613f-4624-9eaa-29fd11f02cfd","Type":"ContainerStarted","Data":"5d1de2b8eb434b4bc2c9112cdc801c6ad7e2dffd2e42b6dc034aae5a698c28da"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.119480 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-h5tlv" podStartSLOduration=120.119464746 podStartE2EDuration="2m0.119464746s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.114621726 +0000 UTC m=+144.981592113" watchObservedRunningTime="2026-01-23 08:54:18.119464746 +0000 UTC m=+144.986435133" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.121153 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zjw59" podStartSLOduration=121.121144208 podStartE2EDuration="2m1.121144208s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.051619863 +0000 UTC m=+144.918590240" watchObservedRunningTime="2026-01-23 08:54:18.121144208 +0000 UTC m=+144.988114595" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.122940 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" event={"ID":"5fee33cd-f588-4cfe-a9a4-90fcb301bccb","Type":"ContainerStarted","Data":"2e52a3d801c755b3289ebbaf33802059430c998789873dfac765fb7ecbbc363b"} Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.122970 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.124827 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.151519 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.164130 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.166642 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.167837 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.667818756 +0000 UTC m=+145.534789143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.202054 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.267305 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.268950 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.768930594 +0000 UTC m=+145.635900971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.365409 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dd4dm" podStartSLOduration=120.365391627 podStartE2EDuration="2m0.365391627s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.218985265 +0000 UTC m=+145.085955652" watchObservedRunningTime="2026-01-23 08:54:18.365391627 +0000 UTC m=+145.232362014" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.365814 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hgqbd" podStartSLOduration=9.365808638 podStartE2EDuration="9.365808638s" podCreationTimestamp="2026-01-23 08:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.330686556 +0000 UTC m=+145.197656943" watchObservedRunningTime="2026-01-23 08:54:18.365808638 +0000 UTC m=+145.232779025" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.374229 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.374677 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.874655367 +0000 UTC m=+145.741625754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.470398 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" podStartSLOduration=121.470382702 podStartE2EDuration="2m1.470382702s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.459247736 +0000 UTC m=+145.326218133" watchObservedRunningTime="2026-01-23 08:54:18.470382702 +0000 UTC m=+145.337353089" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.475392 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.475592 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.97556211 +0000 UTC m=+145.842532497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.475715 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.476160 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:18.976151745 +0000 UTC m=+145.843122132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.555892 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6jqvj" podStartSLOduration=120.555876513 podStartE2EDuration="2m0.555876513s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.534125883 +0000 UTC m=+145.401096290" watchObservedRunningTime="2026-01-23 08:54:18.555876513 +0000 UTC m=+145.422846900" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.577521 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.577717 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.077685994 +0000 UTC m=+145.944656381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.577851 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.578184 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.078175336 +0000 UTC m=+145.945145723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.619221 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-krd2c" podStartSLOduration=120.619188643 podStartE2EDuration="2m0.619188643s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.615188894 +0000 UTC m=+145.482159281" watchObservedRunningTime="2026-01-23 08:54:18.619188643 +0000 UTC m=+145.486159030" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.641393 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:18 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:18 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:18 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.641442 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.679226 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.679538 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.17952574 +0000 UTC m=+146.046496117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.779876 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8bmp9" podStartSLOduration=120.779858539 podStartE2EDuration="2m0.779858539s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.777327596 +0000 UTC m=+145.644297983" watchObservedRunningTime="2026-01-23 08:54:18.779858539 +0000 UTC m=+145.646828926" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.780180 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.780559 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.280547596 +0000 UTC m=+146.147517983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.801452 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c49bm" podStartSLOduration=120.801434394 podStartE2EDuration="2m0.801434394s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.800546602 +0000 UTC m=+145.667516989" watchObservedRunningTime="2026-01-23 08:54:18.801434394 +0000 UTC m=+145.668404781" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.848231 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h2gr8" podStartSLOduration=120.848199554 podStartE2EDuration="2m0.848199554s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.846155984 +0000 UTC m=+145.713126371" watchObservedRunningTime="2026-01-23 08:54:18.848199554 +0000 UTC m=+145.715169941" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.875810 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gwvhs" podStartSLOduration=120.875789279 podStartE2EDuration="2m0.875789279s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.875507792 +0000 UTC m=+145.742478179" watchObservedRunningTime="2026-01-23 08:54:18.875789279 +0000 UTC m=+145.742759676" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.881093 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.881505 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.38149127 +0000 UTC m=+146.248461647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.924945 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-clf44" podStartSLOduration=120.924930588 podStartE2EDuration="2m0.924930588s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.920436046 +0000 UTC m=+145.787406433" watchObservedRunningTime="2026-01-23 08:54:18.924930588 +0000 UTC m=+145.791900975" Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.982696 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:18 crc kubenswrapper[4899]: E0123 08:54:18.983047 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.483030399 +0000 UTC m=+146.350000786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:18 crc kubenswrapper[4899]: I0123 08:54:18.998759 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5wwg" podStartSLOduration=120.998743649 podStartE2EDuration="2m0.998743649s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:18.997237342 +0000 UTC m=+145.864207739" watchObservedRunningTime="2026-01-23 08:54:18.998743649 +0000 UTC m=+145.865714026" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.029564 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" podStartSLOduration=121.029545763 podStartE2EDuration="2m1.029545763s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:19.022894808 +0000 UTC m=+145.889865195" watchObservedRunningTime="2026-01-23 08:54:19.029545763 +0000 UTC m=+145.896516150" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.083814 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.084267 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.58425076 +0000 UTC m=+146.451221147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.132373 4899 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nxzpb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.132430 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" podUID="80c7e7bf-f135-4728-b27b-0b23b8093b07" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.139514 4899 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ctjz9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.139860 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.144626 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-28lqh" podStartSLOduration=121.144609018 podStartE2EDuration="2m1.144609018s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:19.099478678 +0000 UTC m=+145.966449065" watchObservedRunningTime="2026-01-23 08:54:19.144609018 +0000 UTC m=+146.011579405" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.144769 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-tbwm2" podStartSLOduration=121.144766342 podStartE2EDuration="2m1.144766342s" podCreationTimestamp="2026-01-23 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:19.143336436 +0000 UTC m=+146.010306823" watchObservedRunningTime="2026-01-23 08:54:19.144766342 +0000 UTC m=+146.011736729" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.187567 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.187931 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.687918512 +0000 UTC m=+146.554888899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.216231 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" event={"ID":"48fd8d73-d00c-4512-9a1c-5669687a6f5c","Type":"ContainerStarted","Data":"f0c0c92f1a45fba3ed06b7fb71ee2e710abd3e3a26da2ed9deb763234a48ebe5"} Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.216269 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" event={"ID":"48fd8d73-d00c-4512-9a1c-5669687a6f5c","Type":"ContainerStarted","Data":"c5c6ef12583b21a6a81b76c682cffc880f65997ddc8a222e046bbda9c37678da"} Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.288785 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.291024 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.791004178 +0000 UTC m=+146.657974575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.393992 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.394330 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.894318981 +0000 UTC m=+146.761289368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.495128 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.496200 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.996156658 +0000 UTC m=+146.863127055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.496431 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.496784 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:19.996773333 +0000 UTC m=+146.863743720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.597038 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.597253 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.097220645 +0000 UTC m=+146.964191042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.597431 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.597778 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.097767618 +0000 UTC m=+146.964738065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.638685 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:19 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:19 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:19 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.638739 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.698056 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.698424 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.198407685 +0000 UTC m=+147.065378072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.799367 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.799744 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.299733629 +0000 UTC m=+147.166704006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.851029 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqzln" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.901044 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.901216 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.401181165 +0000 UTC m=+147.268151542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.901350 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:19 crc kubenswrapper[4899]: E0123 08:54:19.901672 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.401665287 +0000 UTC m=+147.268635674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.918365 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.957824 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.958817 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.979227 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 23 08:54:19 crc kubenswrapper[4899]: I0123 08:54:19.982333 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.003754 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.004053 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.004262 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.504238872 +0000 UTC m=+147.371209259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.004461 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74nhw\" (UniqueName: \"kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.004610 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.105969 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.106042 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.106106 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74nhw\" (UniqueName: \"kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.106134 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.106529 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.106686 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.106861 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.606842057 +0000 UTC m=+147.473812444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.210546 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74nhw\" (UniqueName: \"kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw\") pod \"certified-operators-zgmpt\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.211253 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.211629 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.711616227 +0000 UTC m=+147.578586614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.214612 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.215758 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.220833 4899 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nxzpb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.220887 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" podUID="80c7e7bf-f135-4728-b27b-0b23b8093b07" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.225093 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" event={"ID":"48fd8d73-d00c-4512-9a1c-5669687a6f5c","Type":"ContainerStarted","Data":"68fb161f968fc94f82ddc4a504516ff0b9013520b4473cead95f879d203aa719"} Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.226407 4899 generic.go:334] "Generic (PLEG): container finished" podID="e5eca0f4-a14b-436b-8d5e-045d7214deee" containerID="62762c4c938b183f7cad91a1b3b8c43e5d30af97182033f5f3a5c4b9e1dabedc" exitCode=0 Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.226584 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" event={"ID":"e5eca0f4-a14b-436b-8d5e-045d7214deee","Type":"ContainerDied","Data":"62762c4c938b183f7cad91a1b3b8c43e5d30af97182033f5f3a5c4b9e1dabedc"} Jan 23 08:54:20 crc kubenswrapper[4899]: W0123 08:54:20.230467 4899 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.230627 4899 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.286896 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.288028 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.312412 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.312468 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prgrc\" (UniqueName: \"kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.312602 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.312635 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.314258 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.814246083 +0000 UTC m=+147.681216470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.329079 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.330277 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.424841 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425323 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prgrc\" (UniqueName: \"kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425459 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78hst\" (UniqueName: \"kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425600 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425745 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425869 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.425979 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.426188 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:20.926167199 +0000 UTC m=+147.793137586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.426516 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.426954 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.427255 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.498557 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prgrc\" (UniqueName: \"kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc\") pod \"community-operators-zcjjh\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.520798 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.521792 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.527074 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.527416 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78hst\" (UniqueName: \"kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.527510 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.527536 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.027521054 +0000 UTC m=+147.894491441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.527688 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.528224 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.528250 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.540827 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.578317 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78hst\" (UniqueName: \"kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst\") pod \"certified-operators-9f6g6\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.629102 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.629344 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.629399 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvpl\" (UniqueName: \"kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.629458 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.629568 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.129553665 +0000 UTC m=+147.996524052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.652654 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:20 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:20 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:20 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.652705 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.711674 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.730633 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.730699 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.730740 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.730761 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvpl\" (UniqueName: \"kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.731519 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.731746 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.732089 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.232064428 +0000 UTC m=+148.099034885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.753393 4899 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.767673 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvpl\" (UniqueName: \"kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl\") pod \"community-operators-vdpv6\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.831735 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.832242 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.332226142 +0000 UTC m=+148.199196539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.839263 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:54:20 crc kubenswrapper[4899]: I0123 08:54:20.932847 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:20 crc kubenswrapper[4899]: E0123 08:54:20.933222 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.433193157 +0000 UTC m=+148.300163544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.033838 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.034063 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.534031439 +0000 UTC m=+148.401001846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.034150 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.034579 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.534564202 +0000 UTC m=+148.401534589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.135812 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.136022 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.635993248 +0000 UTC m=+148.502963635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.136188 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.136640 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.636631854 +0000 UTC m=+148.503602241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.224296 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.227274 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.228666 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.239143 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" event={"ID":"48fd8d73-d00c-4512-9a1c-5669687a6f5c","Type":"ContainerStarted","Data":"e23f9f47ea90aad0f1c1ac44927ef54d2a2b39d5c66fcdb60056cd52cf1c6b3c"} Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.240589 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.241075 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.741056705 +0000 UTC m=+148.608027102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.244876 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerStarted","Data":"cc9034bd894bd51eb5b2440a3bca949000815fd9243f4a74edf2d322064bf982"} Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.267372 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cbtmt" podStartSLOduration=12.267352137 podStartE2EDuration="12.267352137s" podCreationTimestamp="2026-01-23 08:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:21.265629354 +0000 UTC m=+148.132599761" watchObservedRunningTime="2026-01-23 08:54:21.267352137 +0000 UTC m=+148.134322524" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.298635 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.342967 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.343418 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.843403854 +0000 UTC m=+148.710374241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444124 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.444322 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.944289556 +0000 UTC m=+148.811259943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444408 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444452 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444519 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444556 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.444602 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.444862 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:21.94484544 +0000 UTC m=+148.811815827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.449636 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.449784 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.451305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.457961 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.507112 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:54:21 crc kubenswrapper[4899]: W0123 08:54:21.514412 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode74c9e8d_7ce9_44fc_b33d_b6fc33d31a50.slice/crio-027879a8c5b62bb13d9ccdcc482529e19585bbbd26b715f8f85d8e6c7cce631e WatchSource:0}: Error finding container 027879a8c5b62bb13d9ccdcc482529e19585bbbd26b715f8f85d8e6c7cce631e: Status 404 returned error can't find the container with id 027879a8c5b62bb13d9ccdcc482529e19585bbbd26b715f8f85d8e6c7cce631e Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.536031 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.545136 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.545201 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume\") pod \"e5eca0f4-a14b-436b-8d5e-045d7214deee\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.545312 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:22.045284722 +0000 UTC m=+148.912255119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.547296 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume\") pod \"e5eca0f4-a14b-436b-8d5e-045d7214deee\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.547354 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgp2x\" (UniqueName: \"kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x\") pod \"e5eca0f4-a14b-436b-8d5e-045d7214deee\" (UID: \"e5eca0f4-a14b-436b-8d5e-045d7214deee\") " Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.547662 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.548013 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume" (OuterVolumeSpecName: "config-volume") pod "e5eca0f4-a14b-436b-8d5e-045d7214deee" (UID: "e5eca0f4-a14b-436b-8d5e-045d7214deee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.548024 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-23 08:54:22.048011769 +0000 UTC m=+148.914982156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l2tzs" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.553010 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.553737 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x" (OuterVolumeSpecName: "kube-api-access-lgp2x") pod "e5eca0f4-a14b-436b-8d5e-045d7214deee" (UID: "e5eca0f4-a14b-436b-8d5e-045d7214deee"). InnerVolumeSpecName "kube-api-access-lgp2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.554544 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e5eca0f4-a14b-436b-8d5e-045d7214deee" (UID: "e5eca0f4-a14b-436b-8d5e-045d7214deee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.590409 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.650286 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:21 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:21 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:21 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.650341 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.650865 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.651292 4899 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5eca0f4-a14b-436b-8d5e-045d7214deee-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.651307 4899 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5eca0f4-a14b-436b-8d5e-045d7214deee-config-volume\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.651320 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgp2x\" (UniqueName: \"kubernetes.io/projected/e5eca0f4-a14b-436b-8d5e-045d7214deee-kube-api-access-lgp2x\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.651399 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-23 08:54:22.151379964 +0000 UTC m=+149.018350351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.693770 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.733360 4899 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-23T08:54:20.753416477Z","Handler":null,"Name":""} Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.736520 4899 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.736556 4899 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.755987 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: W0123 08:54:21.775588 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-938257de1cbc40ea1dd0591f46999048eb40544b00665c4a297f6e205c10f28f WatchSource:0}: Error finding container 938257de1cbc40ea1dd0591f46999048eb40544b00665c4a297f6e205c10f28f: Status 404 returned error can't find the container with id 938257de1cbc40ea1dd0591f46999048eb40544b00665c4a297f6e205c10f28f Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.798487 4899 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.798564 4899 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.815731 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 23 08:54:21 crc kubenswrapper[4899]: E0123 08:54:21.816017 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5eca0f4-a14b-436b-8d5e-045d7214deee" containerName="collect-profiles" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.816094 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5eca0f4-a14b-436b-8d5e-045d7214deee" containerName="collect-profiles" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.816257 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5eca0f4-a14b-436b-8d5e-045d7214deee" containerName="collect-profiles" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.817006 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.820112 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.820770 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.823031 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 23 08:54:21 crc kubenswrapper[4899]: W0123 08:54:21.824905 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b7982c7824d2bca6ef4eb2ee8e7c3240f591a11ecebb6c85d252bbd5969ba6eb WatchSource:0}: Error finding container b7982c7824d2bca6ef4eb2ee8e7c3240f591a11ecebb6c85d252bbd5969ba6eb: Status 404 returned error can't find the container with id b7982c7824d2bca6ef4eb2ee8e7c3240f591a11ecebb6c85d252bbd5969ba6eb Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.836391 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l2tzs\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.843079 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.847958 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.848056 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.858104 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.858134 4899 patch_prober.go:28] interesting pod/apiserver-76f77b778f-k4cjk container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]log ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]etcd ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/generic-apiserver-start-informers ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/max-in-flight-filter ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 23 08:54:21 crc kubenswrapper[4899]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 23 08:54:21 crc kubenswrapper[4899]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/project.openshift.io-projectcache ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-startinformers ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 23 08:54:21 crc kubenswrapper[4899]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 23 08:54:21 crc kubenswrapper[4899]: livez check failed Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.858185 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" podUID="986b43bc-5ff7-486a-8539-f7568bc468f1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.858282 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.858317 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.869937 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.874468 4899 patch_prober.go:28] interesting pod/downloads-7954f5f757-r6525 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.874471 4899 patch_prober.go:28] interesting pod/downloads-7954f5f757-r6525 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.874519 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r6525" podUID="0be4b82f-6b52-486a-8f68-88a6cd3a8317" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.874529 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r6525" podUID="0be4b82f-6b52-486a-8f68-88a6cd3a8317" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.903583 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.904367 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.907028 4899 patch_prober.go:28] interesting pod/console-f9d7485db-n7s8j container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.907080 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-n7s8j" podUID="27183ff6-88f1-42ef-9d0c-bbd4da6d2441" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.919477 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.933626 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.933669 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.939389 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.964860 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.965411 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.965802 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:21 crc kubenswrapper[4899]: I0123 08:54:21.989630 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:22 crc kubenswrapper[4899]: W0123 08:54:22.104104 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b098e068eeb42f0a172846f7ceec992aa71958fc213e4e63b87814c49b242788 WatchSource:0}: Error finding container b098e068eeb42f0a172846f7ceec992aa71958fc213e4e63b87814c49b242788: Status 404 returned error can't find the container with id b098e068eeb42f0a172846f7ceec992aa71958fc213e4e63b87814c49b242788 Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.104109 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.105673 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.108256 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.122501 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.135037 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.139246 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.168190 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcb4b\" (UniqueName: \"kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.169953 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.169988 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.232439 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.260670 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerStarted","Data":"7543d4e953b5fea41c893043f53e7cdd69405a730e9b8119a2a6effd47601e9d"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.264138 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b7982c7824d2bca6ef4eb2ee8e7c3240f591a11ecebb6c85d252bbd5969ba6eb"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.265458 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" event={"ID":"e5eca0f4-a14b-436b-8d5e-045d7214deee","Type":"ContainerDied","Data":"9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.265479 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9063d6e5ef94d5827644000428f8f58349afc5da2a32f233ff11e85555dde604" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.265621 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485965-g8clk" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.270372 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerStarted","Data":"a3d567bda0f8df27a4d98123609afda546e083c5575e543aa78f67305f266f44"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.270932 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.271079 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcb4b\" (UniqueName: \"kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.271142 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.271524 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.271631 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.290777 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b098e068eeb42f0a172846f7ceec992aa71958fc213e4e63b87814c49b242788"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.294108 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerStarted","Data":"027879a8c5b62bb13d9ccdcc482529e19585bbbd26b715f8f85d8e6c7cce631e"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.301672 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"938257de1cbc40ea1dd0591f46999048eb40544b00665c4a297f6e205c10f28f"} Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.314371 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcb4b\" (UniqueName: \"kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b\") pod \"redhat-marketplace-h4ckf\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.316506 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5bwgh" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.438541 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.464492 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.503680 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.504998 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.520869 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.581343 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x47fq\" (UniqueName: \"kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.581401 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.581431 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.594381 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:54:22 crc kubenswrapper[4899]: W0123 08:54:22.613676 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc1b652a_534c_414e_86b3_00b89a82c21e.slice/crio-b616b35e36af780631ba6464a79f5cbbe16c4c3a725134ade56d987901b6afca WatchSource:0}: Error finding container b616b35e36af780631ba6464a79f5cbbe16c4c3a725134ade56d987901b6afca: Status 404 returned error can't find the container with id b616b35e36af780631ba6464a79f5cbbe16c4c3a725134ade56d987901b6afca Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.634388 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.638322 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:22 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:22 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:22 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.638410 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.682286 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x47fq\" (UniqueName: \"kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.684120 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.684165 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.684657 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.685066 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.701799 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x47fq\" (UniqueName: \"kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq\") pod \"redhat-marketplace-cj8l9\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.794085 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.844127 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxzpb" Jan 23 08:54:22 crc kubenswrapper[4899]: I0123 08:54:22.858795 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.126428 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.132091 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.134220 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.135082 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.140738 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.191415 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.191816 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sw57\" (UniqueName: \"kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.191847 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.292816 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.292880 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sw57\" (UniqueName: \"kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.292904 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.293780 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.293795 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.308416 4899 generic.go:334] "Generic (PLEG): container finished" podID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerID="892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e" exitCode=0 Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.308590 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerDied","Data":"892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.310403 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" event={"ID":"bc1b652a-534c-414e-86b3-00b89a82c21e","Type":"ContainerStarted","Data":"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.310426 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" event={"ID":"bc1b652a-534c-414e-86b3-00b89a82c21e","Type":"ContainerStarted","Data":"b616b35e36af780631ba6464a79f5cbbe16c4c3a725134ade56d987901b6afca"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.310721 4899 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.311029 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.313943 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"02d159da13e803bedcea1f7b7a053dbe23b025b2d9e67db95dbe9a8e0f68ec0e"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.321746 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b35e170ad6e450c8c4eed8ca2371d6685b1204c6fbe39ac4a164a0dda1c40ade"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.321962 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.322379 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sw57\" (UniqueName: \"kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57\") pod \"redhat-operators-rstzc\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.327158 4899 generic.go:334] "Generic (PLEG): container finished" podID="56163971-32c1-44ab-9d10-d884891a2440" containerID="68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d" exitCode=0 Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.327319 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerDied","Data":"68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.338066 4899 generic.go:334] "Generic (PLEG): container finished" podID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerID="84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70" exitCode=0 Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.338142 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerDied","Data":"84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.348798 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5c62536d492f0b510d226e0ac44dce8c9ba0498e2b5b8ac11c300a64d0c7acd4"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.356202 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" podStartSLOduration=126.356185755 podStartE2EDuration="2m6.356185755s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:23.35477056 +0000 UTC m=+150.221740957" watchObservedRunningTime="2026-01-23 08:54:23.356185755 +0000 UTC m=+150.223156142" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.358533 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerStarted","Data":"ec3e3726fe0729601d421884bfe6674eb36067bb5e67573ecf4f870e3ed8eb81"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.368521 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2d27feb3-ad8c-4129-b09e-c6ad37765a31","Type":"ContainerStarted","Data":"6d9662399440916b8417a1c336f4d791570307576b7e820182c2caeb319328e9"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.368712 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2d27feb3-ad8c-4129-b09e-c6ad37765a31","Type":"ContainerStarted","Data":"a60a619e3bde66fb8e6e6954c678366c65e91c7d7a6182d5aa945e364f7c9b54"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.377623 4899 generic.go:334] "Generic (PLEG): container finished" podID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerID="32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0" exitCode=0 Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.377732 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerDied","Data":"32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.391982 4899 generic.go:334] "Generic (PLEG): container finished" podID="a124473f-1d9e-40ff-99bb-5413ce650662" containerID="4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9" exitCode=0 Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.393241 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerDied","Data":"4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.393283 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerStarted","Data":"92e511a5f80f33424bfbaa8e7e26219fb0464a98551606dd1cf2f42e92f7bbb4"} Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.477962 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.508001 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.509041 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.524646 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.529290 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.603242 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.603324 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q99w\" (UniqueName: \"kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.603351 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.605702 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.605689134 podStartE2EDuration="2.605689134s" podCreationTimestamp="2026-01-23 08:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:54:23.604442854 +0000 UTC m=+150.471413241" watchObservedRunningTime="2026-01-23 08:54:23.605689134 +0000 UTC m=+150.472659511" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.640565 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:23 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:23 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:23 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.640957 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.704405 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.704801 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.704848 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q99w\" (UniqueName: \"kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.705317 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.705527 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.758673 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q99w\" (UniqueName: \"kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w\") pod \"redhat-operators-mdwxw\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.814819 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:54:23 crc kubenswrapper[4899]: I0123 08:54:23.906841 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.116161 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.399115 4899 generic.go:334] "Generic (PLEG): container finished" podID="d641b177-76eb-4bb8-bd36-238f8907701d" containerID="621bc9c552313230bbe38a5a63ff4d1f57d47c07b366f51a05e83eccbd115afa" exitCode=0 Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.399218 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerDied","Data":"621bc9c552313230bbe38a5a63ff4d1f57d47c07b366f51a05e83eccbd115afa"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.400849 4899 generic.go:334] "Generic (PLEG): container finished" podID="2d27feb3-ad8c-4129-b09e-c6ad37765a31" containerID="6d9662399440916b8417a1c336f4d791570307576b7e820182c2caeb319328e9" exitCode=0 Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.400874 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2d27feb3-ad8c-4129-b09e-c6ad37765a31","Type":"ContainerDied","Data":"6d9662399440916b8417a1c336f4d791570307576b7e820182c2caeb319328e9"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.402357 4899 generic.go:334] "Generic (PLEG): container finished" podID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerID="8f5087e1f8c7b43f0a88c063ea996d3969b9f0990b40e3e154fc9446c528189e" exitCode=0 Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.402433 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerDied","Data":"8f5087e1f8c7b43f0a88c063ea996d3969b9f0990b40e3e154fc9446c528189e"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.402505 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerStarted","Data":"5e093e40763c7cb775bb91a6ba3b41886fb2fb5f323beea87c211bee930c53ed"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.403717 4899 generic.go:334] "Generic (PLEG): container finished" podID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerID="f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534" exitCode=0 Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.403810 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerDied","Data":"f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.403841 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerStarted","Data":"ade66961204c8cc0562aa5b2cc1258f4b98c3204b4f460536ed9316735f7978c"} Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.639151 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:24 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:24 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:24 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:24 crc kubenswrapper[4899]: I0123 08:54:24.639227 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.451535 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.452998 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.455655 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.466187 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.466468 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.537913 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.537980 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.615964 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.637384 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:25 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:25 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:25 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.637474 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.638598 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir\") pod \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.638681 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access\") pod \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\" (UID: \"2d27feb3-ad8c-4129-b09e-c6ad37765a31\") " Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.638717 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2d27feb3-ad8c-4129-b09e-c6ad37765a31" (UID: "2d27feb3-ad8c-4129-b09e-c6ad37765a31"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.638965 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.638999 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.639074 4899 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.639127 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.647573 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2d27feb3-ad8c-4129-b09e-c6ad37765a31" (UID: "2d27feb3-ad8c-4129-b09e-c6ad37765a31"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.655752 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.740345 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d27feb3-ad8c-4129-b09e-c6ad37765a31-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:25 crc kubenswrapper[4899]: I0123 08:54:25.774374 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.045001 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 23 08:54:26 crc kubenswrapper[4899]: W0123 08:54:26.057096 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2c6ddd7d_3fb9_4316_803a_1de8c135a82f.slice/crio-00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e WatchSource:0}: Error finding container 00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e: Status 404 returned error can't find the container with id 00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.418764 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2d27feb3-ad8c-4129-b09e-c6ad37765a31","Type":"ContainerDied","Data":"a60a619e3bde66fb8e6e6954c678366c65e91c7d7a6182d5aa945e364f7c9b54"} Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.418899 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60a619e3bde66fb8e6e6954c678366c65e91c7d7a6182d5aa945e364f7c9b54" Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.418823 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.420741 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2c6ddd7d-3fb9-4316-803a-1de8c135a82f","Type":"ContainerStarted","Data":"00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e"} Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.638982 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:26 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:26 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:26 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.639357 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.855770 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:26 crc kubenswrapper[4899]: I0123 08:54:26.860691 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-k4cjk" Jan 23 08:54:27 crc kubenswrapper[4899]: I0123 08:54:27.430078 4899 generic.go:334] "Generic (PLEG): container finished" podID="2c6ddd7d-3fb9-4316-803a-1de8c135a82f" containerID="a1aece2297df750edac27eb0d61bcef7000200d377d3ce37ec6bc7ff58c4b9fa" exitCode=0 Jan 23 08:54:27 crc kubenswrapper[4899]: I0123 08:54:27.430972 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2c6ddd7d-3fb9-4316-803a-1de8c135a82f","Type":"ContainerDied","Data":"a1aece2297df750edac27eb0d61bcef7000200d377d3ce37ec6bc7ff58c4b9fa"} Jan 23 08:54:27 crc kubenswrapper[4899]: I0123 08:54:27.647814 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:27 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:27 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:27 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:27 crc kubenswrapper[4899]: I0123 08:54:27.649602 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:27 crc kubenswrapper[4899]: I0123 08:54:27.906118 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hgqbd" Jan 23 08:54:28 crc kubenswrapper[4899]: I0123 08:54:28.647124 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:28 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:28 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:28 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:28 crc kubenswrapper[4899]: I0123 08:54:28.647220 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:29 crc kubenswrapper[4899]: I0123 08:54:29.637154 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:29 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:29 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:29 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:29 crc kubenswrapper[4899]: I0123 08:54:29.637327 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:30 crc kubenswrapper[4899]: I0123 08:54:30.636450 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:30 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:30 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:30 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:30 crc kubenswrapper[4899]: I0123 08:54:30.636509 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.637252 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:31 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:31 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:31 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.637534 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.686042 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.686098 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.881137 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-r6525" Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.914443 4899 patch_prober.go:28] interesting pod/console-f9d7485db-n7s8j container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Jan 23 08:54:31 crc kubenswrapper[4899]: I0123 08:54:31.914512 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-n7s8j" podUID="27183ff6-88f1-42ef-9d0c-bbd4da6d2441" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Jan 23 08:54:32 crc kubenswrapper[4899]: I0123 08:54:32.637201 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:32 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:32 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:32 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:32 crc kubenswrapper[4899]: I0123 08:54:32.637281 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:33 crc kubenswrapper[4899]: I0123 08:54:33.637595 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:33 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:33 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:33 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:33 crc kubenswrapper[4899]: I0123 08:54:33.637647 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.460385 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.491986 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2c6ddd7d-3fb9-4316-803a-1de8c135a82f","Type":"ContainerDied","Data":"00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e"} Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492023 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00ffd580899d5cb2c210ddf2ea83e66334574d581a626be7a11ed05566efba2e" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492071 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492499 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access\") pod \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492547 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir\") pod \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\" (UID: \"2c6ddd7d-3fb9-4316-803a-1de8c135a82f\") " Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492696 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2c6ddd7d-3fb9-4316-803a-1de8c135a82f" (UID: "2c6ddd7d-3fb9-4316-803a-1de8c135a82f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.492818 4899 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.499046 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2c6ddd7d-3fb9-4316-803a-1de8c135a82f" (UID: "2c6ddd7d-3fb9-4316-803a-1de8c135a82f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.593711 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c6ddd7d-3fb9-4316-803a-1de8c135a82f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.637060 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:34 crc kubenswrapper[4899]: [-]has-synced failed: reason withheld Jan 23 08:54:34 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:34 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:34 crc kubenswrapper[4899]: I0123 08:54:34.637123 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:35 crc kubenswrapper[4899]: I0123 08:54:35.637871 4899 patch_prober.go:28] interesting pod/router-default-5444994796-hszft container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 23 08:54:35 crc kubenswrapper[4899]: [+]has-synced ok Jan 23 08:54:35 crc kubenswrapper[4899]: [+]process-running ok Jan 23 08:54:35 crc kubenswrapper[4899]: healthz check failed Jan 23 08:54:35 crc kubenswrapper[4899]: I0123 08:54:35.638195 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hszft" podUID="6ea85a4b-f44f-4500-94d8-99c116ab4e43" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 23 08:54:36 crc kubenswrapper[4899]: I0123 08:54:36.639163 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:36 crc kubenswrapper[4899]: I0123 08:54:36.641786 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-hszft" Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.077155 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.100087 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb844e06-c862-406b-8ecb-36b682d2d58b-metrics-certs\") pod \"network-metrics-daemon-kxcxw\" (UID: \"bb844e06-c862-406b-8ecb-36b682d2d58b\") " pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.158826 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxcxw" Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.797020 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.797545 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" containerID="cri-o://24d64bb049869912d764aa8b174fd919f140511fb113c25967481188dba33401" gracePeriod=30 Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.820619 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:40 crc kubenswrapper[4899]: I0123 08:54:40.820860 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" containerID="cri-o://05a5fc672ac3db17e00405398b1a20edb90d73fcc4c9a87b9d33766b6250d960" gracePeriod=30 Jan 23 08:54:41 crc kubenswrapper[4899]: I0123 08:54:41.928266 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.126694 4899 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vfz69 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.126803 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.300196 4899 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2xtw6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.300285 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.404969 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:42 crc kubenswrapper[4899]: I0123 08:54:42.409735 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-n7s8j" Jan 23 08:54:45 crc kubenswrapper[4899]: I0123 08:54:45.552762 4899 generic.go:334] "Generic (PLEG): container finished" podID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerID="24d64bb049869912d764aa8b174fd919f140511fb113c25967481188dba33401" exitCode=0 Jan 23 08:54:45 crc kubenswrapper[4899]: I0123 08:54:45.552810 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" event={"ID":"9f959534-2339-44d2-82dc-82f6c8e8a893","Type":"ContainerDied","Data":"24d64bb049869912d764aa8b174fd919f140511fb113c25967481188dba33401"} Jan 23 08:54:47 crc kubenswrapper[4899]: I0123 08:54:47.571411 4899 generic.go:334] "Generic (PLEG): container finished" podID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerID="05a5fc672ac3db17e00405398b1a20edb90d73fcc4c9a87b9d33766b6250d960" exitCode=0 Jan 23 08:54:47 crc kubenswrapper[4899]: I0123 08:54:47.571462 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" event={"ID":"8a08d77e-7f9c-4c41-8d85-95b64e461ea5","Type":"ContainerDied","Data":"05a5fc672ac3db17e00405398b1a20edb90d73fcc4c9a87b9d33766b6250d960"} Jan 23 08:54:49 crc kubenswrapper[4899]: E0123 08:54:49.191673 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2264292725/2\": happened during read: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 23 08:54:49 crc kubenswrapper[4899]: E0123 08:54:49.192483 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-prgrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zcjjh_openshift-marketplace(e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2264292725/2\": happened during read: context canceled" logger="UnhandledError" Jan 23 08:54:49 crc kubenswrapper[4899]: E0123 08:54:49.193679 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage2264292725/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/community-operators-zcjjh" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" Jan 23 08:54:52 crc kubenswrapper[4899]: I0123 08:54:52.126609 4899 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vfz69 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 23 08:54:52 crc kubenswrapper[4899]: I0123 08:54:52.126924 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 23 08:54:52 crc kubenswrapper[4899]: I0123 08:54:52.163805 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ddgpg" Jan 23 08:54:52 crc kubenswrapper[4899]: I0123 08:54:52.873587 4899 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-hhxvh container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:54:52 crc kubenswrapper[4899]: I0123 08:54:52.873955 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhxvh" podUID="433100b8-a3d5-4f8a-8e88-57026c742941" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.248174 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zcjjh" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.300519 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.300622 4899 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2xtw6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.300652 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.333902 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.334150 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334167 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.334215 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d27feb3-ad8c-4129-b09e-c6ad37765a31" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334223 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d27feb3-ad8c-4129-b09e-c6ad37765a31" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.334232 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6ddd7d-3fb9-4316-803a-1de8c135a82f" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334240 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6ddd7d-3fb9-4316-803a-1de8c135a82f" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334386 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" containerName="route-controller-manager" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334402 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d27feb3-ad8c-4129-b09e-c6ad37765a31" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334414 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6ddd7d-3fb9-4316-803a-1de8c135a82f" containerName="pruner" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.334994 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.339861 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.373787 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config\") pod \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.373878 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccz5d\" (UniqueName: \"kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d\") pod \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.373915 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca\") pod \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.373937 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert\") pod \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\" (UID: \"8a08d77e-7f9c-4c41-8d85-95b64e461ea5\") " Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.374122 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.374177 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.374305 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.374359 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2btj\" (UniqueName: \"kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.374755 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca" (OuterVolumeSpecName: "client-ca") pod "8a08d77e-7f9c-4c41-8d85-95b64e461ea5" (UID: "8a08d77e-7f9c-4c41-8d85-95b64e461ea5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.375344 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config" (OuterVolumeSpecName: "config") pod "8a08d77e-7f9c-4c41-8d85-95b64e461ea5" (UID: "8a08d77e-7f9c-4c41-8d85-95b64e461ea5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.378783 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8a08d77e-7f9c-4c41-8d85-95b64e461ea5" (UID: "8a08d77e-7f9c-4c41-8d85-95b64e461ea5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.384169 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d" (OuterVolumeSpecName: "kube-api-access-ccz5d") pod "8a08d77e-7f9c-4c41-8d85-95b64e461ea5" (UID: "8a08d77e-7f9c-4c41-8d85-95b64e461ea5"). InnerVolumeSpecName "kube-api-access-ccz5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.424649 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.425140 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78hst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9f6g6_openshift-marketplace(56163971-32c1-44ab-9d10-d884891a2440): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.426306 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9f6g6" podUID="56163971-32c1-44ab-9d10-d884891a2440" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476337 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2btj\" (UniqueName: \"kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476400 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476434 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476471 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476510 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476521 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476531 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.476540 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccz5d\" (UniqueName: \"kubernetes.io/projected/8a08d77e-7f9c-4c41-8d85-95b64e461ea5-kube-api-access-ccz5d\") on node \"crc\" DevicePath \"\"" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.477605 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.477705 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.492824 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.497147 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2btj\" (UniqueName: \"kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj\") pod \"route-controller-manager-6b469fb7f8-prxvn\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.543860 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage874077828/2\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.544023 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pcb4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-h4ckf_openshift-marketplace(a124473f-1d9e-40ff-99bb-5413ce650662): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage874077828/2\": happened during read: context canceled" logger="UnhandledError" Jan 23 08:54:53 crc kubenswrapper[4899]: E0123 08:54:53.545461 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage874077828/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-marketplace-h4ckf" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.606341 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" event={"ID":"8a08d77e-7f9c-4c41-8d85-95b64e461ea5","Type":"ContainerDied","Data":"6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53"} Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.606405 4899 scope.go:117] "RemoveContainer" containerID="05a5fc672ac3db17e00405398b1a20edb90d73fcc4c9a87b9d33766b6250d960" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.606686 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6" Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.654021 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.656993 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2xtw6"] Jan 23 08:54:53 crc kubenswrapper[4899]: I0123 08:54:53.658576 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:54:54 crc kubenswrapper[4899]: E0123 08:54:54.870278 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:54:55 crc kubenswrapper[4899]: I0123 08:54:55.523036 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a08d77e-7f9c-4c41-8d85-95b64e461ea5" path="/var/lib/kubelet/pods/8a08d77e-7f9c-4c41-8d85-95b64e461ea5/volumes" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.765680 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-h4ckf" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.765706 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9f6g6" podUID="56163971-32c1-44ab-9d10-d884891a2440" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.782247 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2043982775/2\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.782393 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x47fq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-cj8l9_openshift-marketplace(d641b177-76eb-4bb8-bd36-238f8907701d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2043982775/2\": happened during read: context canceled" logger="UnhandledError" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.783621 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage2043982775/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-marketplace-cj8l9" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.832716 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.833068 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wjvpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vdpv6_openshift-marketplace(19f0a6a2-df38-4b4a-8844-6a550e0d7eca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 08:54:56 crc kubenswrapper[4899]: E0123 08:54:56.834488 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vdpv6" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.203891 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-cj8l9" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.204149 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vdpv6" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.302325 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.302790 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74nhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zgmpt_openshift-marketplace(4625e21a-e288-4ab7-ac0b-b392f35778ac): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.304057 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zgmpt" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.331643 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.331901 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9sw57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rstzc_openshift-marketplace(9c23cd89-e74f-499c-ab6b-fb76e68004b5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.333887 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rstzc" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.577302 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.599148 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.599441 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.599460 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.599579 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" containerName="controller-manager" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.599997 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.619760 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.653475 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerStarted","Data":"c7f0ea4ae1e11d069900075f8709c4a85c4bb939f4e27cddfb8a470044a4705f"} Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.655911 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" event={"ID":"9f959534-2339-44d2-82dc-82f6c8e8a893","Type":"ContainerDied","Data":"ec62ef2ae22386649078b9bb77b74c2fd030dc51953c7563f0fe407bc19b4163"} Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.656018 4899 scope.go:117] "RemoveContainer" containerID="24d64bb049869912d764aa8b174fd919f140511fb113c25967481188dba33401" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.656262 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vfz69" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.656524 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zgmpt" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" Jan 23 08:55:00 crc kubenswrapper[4899]: E0123 08:55:00.660083 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rstzc" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671180 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert\") pod \"9f959534-2339-44d2-82dc-82f6c8e8a893\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671314 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhqnr\" (UniqueName: \"kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr\") pod \"9f959534-2339-44d2-82dc-82f6c8e8a893\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671342 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca\") pod \"9f959534-2339-44d2-82dc-82f6c8e8a893\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671367 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config\") pod \"9f959534-2339-44d2-82dc-82f6c8e8a893\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671396 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles\") pod \"9f959534-2339-44d2-82dc-82f6c8e8a893\" (UID: \"9f959534-2339-44d2-82dc-82f6c8e8a893\") " Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671545 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671589 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671611 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671628 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghbms\" (UniqueName: \"kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.671649 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.672337 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9f959534-2339-44d2-82dc-82f6c8e8a893" (UID: "9f959534-2339-44d2-82dc-82f6c8e8a893"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.672742 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config" (OuterVolumeSpecName: "config") pod "9f959534-2339-44d2-82dc-82f6c8e8a893" (UID: "9f959534-2339-44d2-82dc-82f6c8e8a893"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.673714 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca" (OuterVolumeSpecName: "client-ca") pod "9f959534-2339-44d2-82dc-82f6c8e8a893" (UID: "9f959534-2339-44d2-82dc-82f6c8e8a893"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.679712 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9f959534-2339-44d2-82dc-82f6c8e8a893" (UID: "9f959534-2339-44d2-82dc-82f6c8e8a893"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.681911 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr" (OuterVolumeSpecName: "kube-api-access-hhqnr") pod "9f959534-2339-44d2-82dc-82f6c8e8a893" (UID: "9f959534-2339-44d2-82dc-82f6c8e8a893"). InnerVolumeSpecName "kube-api-access-hhqnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.695515 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kxcxw"] Jan 23 08:55:00 crc kubenswrapper[4899]: W0123 08:55:00.703690 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb844e06_c862_406b_8ecb_36b682d2d58b.slice/crio-0390b6596b6a64ceb4ee50335295bbfcabdaa90281dd7110c2029fbb5c1a3bfc WatchSource:0}: Error finding container 0390b6596b6a64ceb4ee50335295bbfcabdaa90281dd7110c2029fbb5c1a3bfc: Status 404 returned error can't find the container with id 0390b6596b6a64ceb4ee50335295bbfcabdaa90281dd7110c2029fbb5c1a3bfc Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.705734 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:55:00 crc kubenswrapper[4899]: W0123 08:55:00.715881 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b237ea7_1940_40e0_b085_b95914430c07.slice/crio-54dde6771e352ebde617dcccd81f5fe10f19d1f09ed2cf16a2a28bc215c89823 WatchSource:0}: Error finding container 54dde6771e352ebde617dcccd81f5fe10f19d1f09ed2cf16a2a28bc215c89823: Status 404 returned error can't find the container with id 54dde6771e352ebde617dcccd81f5fe10f19d1f09ed2cf16a2a28bc215c89823 Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.773113 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.774017 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.774413 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776155 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776195 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghbms\" (UniqueName: \"kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776263 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776399 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776418 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776429 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f959534-2339-44d2-82dc-82f6c8e8a893-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776440 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f959534-2339-44d2-82dc-82f6c8e8a893-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.776452 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhqnr\" (UniqueName: \"kubernetes.io/projected/9f959534-2339-44d2-82dc-82f6c8e8a893-kube-api-access-hhqnr\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.777456 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.778177 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.781417 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.796081 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghbms\" (UniqueName: \"kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms\") pod \"controller-manager-856dc49d-wlxkk\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.815547 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.817356 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:00 crc kubenswrapper[4899]: I0123 08:55:00.911027 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.034848 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.038077 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vfz69"] Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.128387 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:01 crc kubenswrapper[4899]: W0123 08:55:01.133612 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18231755_e088_405f_be25_b1c372bc5915.slice/crio-964c54bc752fceec1193200cb2d785df5a8bc98a990c7eedf8cc3beb8c0a5d2b WatchSource:0}: Error finding container 964c54bc752fceec1193200cb2d785df5a8bc98a990c7eedf8cc3beb8c0a5d2b: Status 404 returned error can't find the container with id 964c54bc752fceec1193200cb2d785df5a8bc98a990c7eedf8cc3beb8c0a5d2b Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.520608 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f959534-2339-44d2-82dc-82f6c8e8a893" path="/var/lib/kubelet/pods/9f959534-2339-44d2-82dc-82f6c8e8a893/volumes" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.543257 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.665676 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" event={"ID":"bb844e06-c862-406b-8ecb-36b682d2d58b","Type":"ContainerStarted","Data":"c47aef9f454fa6a272dd98a18d96a5aa70462878245940a2b2e944aa7fae1a71"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.665720 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" event={"ID":"bb844e06-c862-406b-8ecb-36b682d2d58b","Type":"ContainerStarted","Data":"46739adc6570647e45e1f74c26b997c3bda4c2687e43798e9b64c1da21ffcdc1"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.665730 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxcxw" event={"ID":"bb844e06-c862-406b-8ecb-36b682d2d58b","Type":"ContainerStarted","Data":"0390b6596b6a64ceb4ee50335295bbfcabdaa90281dd7110c2029fbb5c1a3bfc"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.669394 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" event={"ID":"18231755-e088-405f-be25-b1c372bc5915","Type":"ContainerStarted","Data":"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.669419 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" event={"ID":"18231755-e088-405f-be25-b1c372bc5915","Type":"ContainerStarted","Data":"964c54bc752fceec1193200cb2d785df5a8bc98a990c7eedf8cc3beb8c0a5d2b"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.669479 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" podUID="18231755-e088-405f-be25-b1c372bc5915" containerName="controller-manager" containerID="cri-o://ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1" gracePeriod=30 Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.669617 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.676412 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" event={"ID":"7b237ea7-1940-40e0-b085-b95914430c07","Type":"ContainerStarted","Data":"012a894d90786c626893987162be70e3879e0da7dfd32a3c4d2589d0a77fccdf"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.676465 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" event={"ID":"7b237ea7-1940-40e0-b085-b95914430c07","Type":"ContainerStarted","Data":"54dde6771e352ebde617dcccd81f5fe10f19d1f09ed2cf16a2a28bc215c89823"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.676784 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.679184 4899 generic.go:334] "Generic (PLEG): container finished" podID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerID="c7f0ea4ae1e11d069900075f8709c4a85c4bb939f4e27cddfb8a470044a4705f" exitCode=0 Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.679228 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerDied","Data":"c7f0ea4ae1e11d069900075f8709c4a85c4bb939f4e27cddfb8a470044a4705f"} Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.682732 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.685031 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kxcxw" podStartSLOduration=164.685014195 podStartE2EDuration="2m44.685014195s" podCreationTimestamp="2026-01-23 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:01.682946324 +0000 UTC m=+188.549916711" watchObservedRunningTime="2026-01-23 08:55:01.685014195 +0000 UTC m=+188.551984582" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.685833 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.685885 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.689780 4899 patch_prober.go:28] interesting pod/controller-manager-856dc49d-wlxkk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": EOF" start-of-body= Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.689840 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" podUID="18231755-e088-405f-be25-b1c372bc5915" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": EOF" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.722854 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" podStartSLOduration=21.722838073 podStartE2EDuration="21.722838073s" podCreationTimestamp="2026-01-23 08:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:01.721398188 +0000 UTC m=+188.588368575" watchObservedRunningTime="2026-01-23 08:55:01.722838073 +0000 UTC m=+188.589808460" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.746648 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" podStartSLOduration=21.746628764 podStartE2EDuration="21.746628764s" podCreationTimestamp="2026-01-23 08:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:01.743709401 +0000 UTC m=+188.610679788" watchObservedRunningTime="2026-01-23 08:55:01.746628764 +0000 UTC m=+188.613599171" Jan 23 08:55:01 crc kubenswrapper[4899]: I0123 08:55:01.991354 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.014851 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:02 crc kubenswrapper[4899]: E0123 08:55:02.015067 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18231755-e088-405f-be25-b1c372bc5915" containerName="controller-manager" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.015079 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="18231755-e088-405f-be25-b1c372bc5915" containerName="controller-manager" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.015187 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="18231755-e088-405f-be25-b1c372bc5915" containerName="controller-manager" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.015619 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.024787 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094161 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghbms\" (UniqueName: \"kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms\") pod \"18231755-e088-405f-be25-b1c372bc5915\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094231 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca\") pod \"18231755-e088-405f-be25-b1c372bc5915\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094302 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles\") pod \"18231755-e088-405f-be25-b1c372bc5915\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094330 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config\") pod \"18231755-e088-405f-be25-b1c372bc5915\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094375 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert\") pod \"18231755-e088-405f-be25-b1c372bc5915\" (UID: \"18231755-e088-405f-be25-b1c372bc5915\") " Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094520 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094553 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094572 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094668 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.094759 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg5z6\" (UniqueName: \"kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.095061 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca" (OuterVolumeSpecName: "client-ca") pod "18231755-e088-405f-be25-b1c372bc5915" (UID: "18231755-e088-405f-be25-b1c372bc5915"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.095070 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "18231755-e088-405f-be25-b1c372bc5915" (UID: "18231755-e088-405f-be25-b1c372bc5915"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.095283 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config" (OuterVolumeSpecName: "config") pod "18231755-e088-405f-be25-b1c372bc5915" (UID: "18231755-e088-405f-be25-b1c372bc5915"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.099746 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "18231755-e088-405f-be25-b1c372bc5915" (UID: "18231755-e088-405f-be25-b1c372bc5915"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.100912 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms" (OuterVolumeSpecName: "kube-api-access-ghbms") pod "18231755-e088-405f-be25-b1c372bc5915" (UID: "18231755-e088-405f-be25-b1c372bc5915"). InnerVolumeSpecName "kube-api-access-ghbms". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196446 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196503 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196526 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196591 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196616 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg5z6\" (UniqueName: \"kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196654 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196664 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196673 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18231755-e088-405f-be25-b1c372bc5915-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196681 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghbms\" (UniqueName: \"kubernetes.io/projected/18231755-e088-405f-be25-b1c372bc5915-kube-api-access-ghbms\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.196690 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18231755-e088-405f-be25-b1c372bc5915-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.197811 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.197946 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.198777 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.201934 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.215263 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg5z6\" (UniqueName: \"kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6\") pod \"controller-manager-6d977694b9-px8jv\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.337139 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.451127 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.458426 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.461100 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.461563 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.462238 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.501609 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.501673 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.529883 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:02 crc kubenswrapper[4899]: W0123 08:55:02.535774 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbef2ec06_fe5c_480b_8d84_67dbc30cbcfd.slice/crio-1b9bc415422ebc06efcc5c4dd3debb2cfa862d04071829c1ee5d7040e05b7a83 WatchSource:0}: Error finding container 1b9bc415422ebc06efcc5c4dd3debb2cfa862d04071829c1ee5d7040e05b7a83: Status 404 returned error can't find the container with id 1b9bc415422ebc06efcc5c4dd3debb2cfa862d04071829c1ee5d7040e05b7a83 Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.602608 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.603258 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.603350 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.623305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.687638 4899 generic.go:334] "Generic (PLEG): container finished" podID="18231755-e088-405f-be25-b1c372bc5915" containerID="ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1" exitCode=0 Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.687723 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.687800 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" event={"ID":"18231755-e088-405f-be25-b1c372bc5915","Type":"ContainerDied","Data":"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1"} Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.687836 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856dc49d-wlxkk" event={"ID":"18231755-e088-405f-be25-b1c372bc5915","Type":"ContainerDied","Data":"964c54bc752fceec1193200cb2d785df5a8bc98a990c7eedf8cc3beb8c0a5d2b"} Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.687862 4899 scope.go:117] "RemoveContainer" containerID="ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.689197 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" event={"ID":"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd","Type":"ContainerStarted","Data":"1b9bc415422ebc06efcc5c4dd3debb2cfa862d04071829c1ee5d7040e05b7a83"} Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.689514 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" podUID="7b237ea7-1940-40e0-b085-b95914430c07" containerName="route-controller-manager" containerID="cri-o://012a894d90786c626893987162be70e3879e0da7dfd32a3c4d2589d0a77fccdf" gracePeriod=30 Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.707181 4899 scope.go:117] "RemoveContainer" containerID="ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1" Jan 23 08:55:02 crc kubenswrapper[4899]: E0123 08:55:02.707903 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1\": container with ID starting with ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1 not found: ID does not exist" containerID="ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.707974 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1"} err="failed to get container status \"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1\": rpc error: code = NotFound desc = could not find container \"ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1\": container with ID starting with ec7ac33f5be84f9c0c61ab7acc9601185d8bb9e4ab67cd51e2fb468e1c4b5de1 not found: ID does not exist" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.744866 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.748491 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-856dc49d-wlxkk"] Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.787538 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:02 crc kubenswrapper[4899]: I0123 08:55:02.972337 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 23 08:55:02 crc kubenswrapper[4899]: W0123 08:55:02.980683 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf7dff80b_7ef9_4933_be6e_50bdf4317e27.slice/crio-0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192 WatchSource:0}: Error finding container 0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192: Status 404 returned error can't find the container with id 0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192 Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.522075 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18231755-e088-405f-be25-b1c372bc5915" path="/var/lib/kubelet/pods/18231755-e088-405f-be25-b1c372bc5915/volumes" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.660814 4899 patch_prober.go:28] interesting pod/route-controller-manager-6b469fb7f8-prxvn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.661173 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" podUID="7b237ea7-1940-40e0-b085-b95914430c07" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.701946 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f7dff80b-7ef9-4933-be6e-50bdf4317e27","Type":"ContainerStarted","Data":"80f758a2bfdf9721024215e500ab948793169db14dc30c0ecb2255baedafadbc"} Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.702288 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f7dff80b-7ef9-4933-be6e-50bdf4317e27","Type":"ContainerStarted","Data":"0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192"} Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.713100 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" event={"ID":"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd","Type":"ContainerStarted","Data":"275d6d681c9f4024a801d4b8bacedf350e93ccfa74f8259823ff9b47dcb12b2e"} Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.713271 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.716929 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerStarted","Data":"7bcc3c44821a82e1216c4397f634d0f9d58f889eb4772664c2c5f0a458a28dc7"} Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.719506 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.721095 4899 generic.go:334] "Generic (PLEG): container finished" podID="7b237ea7-1940-40e0-b085-b95914430c07" containerID="012a894d90786c626893987162be70e3879e0da7dfd32a3c4d2589d0a77fccdf" exitCode=0 Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.721143 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" event={"ID":"7b237ea7-1940-40e0-b085-b95914430c07","Type":"ContainerDied","Data":"012a894d90786c626893987162be70e3879e0da7dfd32a3c4d2589d0a77fccdf"} Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.727811 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.7277812 podStartE2EDuration="1.7277812s" podCreationTimestamp="2026-01-23 08:55:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:03.719418883 +0000 UTC m=+190.586389280" watchObservedRunningTime="2026-01-23 08:55:03.7277812 +0000 UTC m=+190.594751597" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.742026 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" podStartSLOduration=3.742009083 podStartE2EDuration="3.742009083s" podCreationTimestamp="2026-01-23 08:55:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:03.73826935 +0000 UTC m=+190.605239757" watchObservedRunningTime="2026-01-23 08:55:03.742009083 +0000 UTC m=+190.608979470" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.766050 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mdwxw" podStartSLOduration=3.005570678 podStartE2EDuration="40.766032139s" podCreationTimestamp="2026-01-23 08:54:23 +0000 UTC" firstStartedPulling="2026-01-23 08:54:24.40366711 +0000 UTC m=+151.270637497" lastFinishedPulling="2026-01-23 08:55:02.164128571 +0000 UTC m=+189.031098958" observedRunningTime="2026-01-23 08:55:03.764959013 +0000 UTC m=+190.631929400" watchObservedRunningTime="2026-01-23 08:55:03.766032139 +0000 UTC m=+190.633002526" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.907462 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:03 crc kubenswrapper[4899]: I0123 08:55:03.907500 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.203796 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.230521 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:04 crc kubenswrapper[4899]: E0123 08:55:04.231041 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b237ea7-1940-40e0-b085-b95914430c07" containerName="route-controller-manager" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.231057 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b237ea7-1940-40e0-b085-b95914430c07" containerName="route-controller-manager" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.231189 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b237ea7-1940-40e0-b085-b95914430c07" containerName="route-controller-manager" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.231667 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.251331 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.322501 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config\") pod \"7b237ea7-1940-40e0-b085-b95914430c07\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.322562 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert\") pod \"7b237ea7-1940-40e0-b085-b95914430c07\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.322639 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca\") pod \"7b237ea7-1940-40e0-b085-b95914430c07\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.322706 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2btj\" (UniqueName: \"kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj\") pod \"7b237ea7-1940-40e0-b085-b95914430c07\" (UID: \"7b237ea7-1940-40e0-b085-b95914430c07\") " Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.323406 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b237ea7-1940-40e0-b085-b95914430c07" (UID: "7b237ea7-1940-40e0-b085-b95914430c07"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.323531 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config" (OuterVolumeSpecName: "config") pod "7b237ea7-1940-40e0-b085-b95914430c07" (UID: "7b237ea7-1940-40e0-b085-b95914430c07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.327946 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj" (OuterVolumeSpecName: "kube-api-access-l2btj") pod "7b237ea7-1940-40e0-b085-b95914430c07" (UID: "7b237ea7-1940-40e0-b085-b95914430c07"). InnerVolumeSpecName "kube-api-access-l2btj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.332341 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b237ea7-1940-40e0-b085-b95914430c07" (UID: "7b237ea7-1940-40e0-b085-b95914430c07"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424317 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skpf9\" (UniqueName: \"kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424384 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424409 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424517 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424636 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424658 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2btj\" (UniqueName: \"kubernetes.io/projected/7b237ea7-1940-40e0-b085-b95914430c07-kube-api-access-l2btj\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424672 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b237ea7-1940-40e0-b085-b95914430c07-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.424684 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b237ea7-1940-40e0-b085-b95914430c07-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.525932 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skpf9\" (UniqueName: \"kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.526009 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.526040 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.526068 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.527054 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.527561 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.532117 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.546075 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skpf9\" (UniqueName: \"kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9\") pod \"route-controller-manager-bb6f8cc6c-9js9h\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.552740 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.731545 4899 generic.go:334] "Generic (PLEG): container finished" podID="f7dff80b-7ef9-4933-be6e-50bdf4317e27" containerID="80f758a2bfdf9721024215e500ab948793169db14dc30c0ecb2255baedafadbc" exitCode=0 Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.731722 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f7dff80b-7ef9-4933-be6e-50bdf4317e27","Type":"ContainerDied","Data":"80f758a2bfdf9721024215e500ab948793169db14dc30c0ecb2255baedafadbc"} Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.733314 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" event={"ID":"7b237ea7-1940-40e0-b085-b95914430c07","Type":"ContainerDied","Data":"54dde6771e352ebde617dcccd81f5fe10f19d1f09ed2cf16a2a28bc215c89823"} Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.733367 4899 scope.go:117] "RemoveContainer" containerID="012a894d90786c626893987162be70e3879e0da7dfd32a3c4d2589d0a77fccdf" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.734137 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn" Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.784424 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.792548 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b469fb7f8-prxvn"] Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.983056 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mdwxw" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="registry-server" probeResult="failure" output=< Jan 23 08:55:04 crc kubenswrapper[4899]: timeout: failed to connect service ":50051" within 1s Jan 23 08:55:04 crc kubenswrapper[4899]: > Jan 23 08:55:04 crc kubenswrapper[4899]: I0123 08:55:04.997584 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:05 crc kubenswrapper[4899]: E0123 08:55:05.012238 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:55:05 crc kubenswrapper[4899]: I0123 08:55:05.522542 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b237ea7-1940-40e0-b085-b95914430c07" path="/var/lib/kubelet/pods/7b237ea7-1940-40e0-b085-b95914430c07/volumes" Jan 23 08:55:05 crc kubenswrapper[4899]: I0123 08:55:05.742335 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" event={"ID":"163c1af8-2bf4-4a45-86fe-39084913618d","Type":"ContainerStarted","Data":"845b24e80e656a0b5e8bbac827b5f90f5e12e0dc0df03ce635185e18f5ea9070"} Jan 23 08:55:05 crc kubenswrapper[4899]: I0123 08:55:05.742409 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" event={"ID":"163c1af8-2bf4-4a45-86fe-39084913618d","Type":"ContainerStarted","Data":"cc12af7bac7f01902c09a1c05f1d42849f2714c073d53dd5ae46b668d6e77d50"} Jan 23 08:55:05 crc kubenswrapper[4899]: I0123 08:55:05.762112 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" podStartSLOduration=5.762093907 podStartE2EDuration="5.762093907s" podCreationTimestamp="2026-01-23 08:55:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:05.759989415 +0000 UTC m=+192.626959812" watchObservedRunningTime="2026-01-23 08:55:05.762093907 +0000 UTC m=+192.629064284" Jan 23 08:55:05 crc kubenswrapper[4899]: I0123 08:55:05.975435 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.148173 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access\") pod \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.148298 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir\") pod \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\" (UID: \"f7dff80b-7ef9-4933-be6e-50bdf4317e27\") " Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.148490 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f7dff80b-7ef9-4933-be6e-50bdf4317e27" (UID: "f7dff80b-7ef9-4933-be6e-50bdf4317e27"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.148657 4899 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.157162 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f7dff80b-7ef9-4933-be6e-50bdf4317e27" (UID: "f7dff80b-7ef9-4933-be6e-50bdf4317e27"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.249538 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7dff80b-7ef9-4933-be6e-50bdf4317e27-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.750010 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.750008 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f7dff80b-7ef9-4933-be6e-50bdf4317e27","Type":"ContainerDied","Data":"0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192"} Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.750487 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0870c60ad849f188fd936269289042bea808763800e6fd683bc149f509af8192" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.750519 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:06 crc kubenswrapper[4899]: I0123 08:55:06.758480 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.453092 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 23 08:55:07 crc kubenswrapper[4899]: E0123 08:55:07.453377 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7dff80b-7ef9-4933-be6e-50bdf4317e27" containerName="pruner" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.453390 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7dff80b-7ef9-4933-be6e-50bdf4317e27" containerName="pruner" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.453501 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7dff80b-7ef9-4933-be6e-50bdf4317e27" containerName="pruner" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.453885 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.455939 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.456877 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.458300 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.463469 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.463537 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.463688 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.565289 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.565362 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.565391 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.565758 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.565793 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.592022 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access\") pod \"installer-9-crc\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.758079 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerStarted","Data":"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d"} Jan 23 08:55:07 crc kubenswrapper[4899]: I0123 08:55:07.769115 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:08 crc kubenswrapper[4899]: I0123 08:55:08.184320 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 23 08:55:08 crc kubenswrapper[4899]: W0123 08:55:08.190157 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod68143dd5_8349_40d1_ac14_9d30e14ad3af.slice/crio-bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe WatchSource:0}: Error finding container bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe: Status 404 returned error can't find the container with id bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe Jan 23 08:55:08 crc kubenswrapper[4899]: I0123 08:55:08.764170 4899 generic.go:334] "Generic (PLEG): container finished" podID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerID="7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d" exitCode=0 Jan 23 08:55:08 crc kubenswrapper[4899]: I0123 08:55:08.764245 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerDied","Data":"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d"} Jan 23 08:55:08 crc kubenswrapper[4899]: I0123 08:55:08.765374 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"68143dd5-8349-40d1-ac14-9d30e14ad3af","Type":"ContainerStarted","Data":"bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe"} Jan 23 08:55:09 crc kubenswrapper[4899]: I0123 08:55:09.776523 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"68143dd5-8349-40d1-ac14-9d30e14ad3af","Type":"ContainerStarted","Data":"ad7fa93087cbb5c11593435e5a9e42f0f67be3e409764682f63ef2401be0b899"} Jan 23 08:55:09 crc kubenswrapper[4899]: I0123 08:55:09.797079 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.797058793 podStartE2EDuration="2.797058793s" podCreationTimestamp="2026-01-23 08:55:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:09.793056834 +0000 UTC m=+196.660027221" watchObservedRunningTime="2026-01-23 08:55:09.797058793 +0000 UTC m=+196.664029180" Jan 23 08:55:10 crc kubenswrapper[4899]: I0123 08:55:10.783649 4899 generic.go:334] "Generic (PLEG): container finished" podID="56163971-32c1-44ab-9d10-d884891a2440" containerID="754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022" exitCode=0 Jan 23 08:55:10 crc kubenswrapper[4899]: I0123 08:55:10.783710 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerDied","Data":"754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022"} Jan 23 08:55:10 crc kubenswrapper[4899]: I0123 08:55:10.790567 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerStarted","Data":"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54"} Jan 23 08:55:10 crc kubenswrapper[4899]: I0123 08:55:10.817470 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zcjjh" podStartSLOduration=4.605045116 podStartE2EDuration="50.817448566s" podCreationTimestamp="2026-01-23 08:54:20 +0000 UTC" firstStartedPulling="2026-01-23 08:54:23.33947266 +0000 UTC m=+150.206443047" lastFinishedPulling="2026-01-23 08:55:09.55187611 +0000 UTC m=+196.418846497" observedRunningTime="2026-01-23 08:55:10.814539094 +0000 UTC m=+197.681509481" watchObservedRunningTime="2026-01-23 08:55:10.817448566 +0000 UTC m=+197.684418953" Jan 23 08:55:11 crc kubenswrapper[4899]: I0123 08:55:11.227642 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:55:11 crc kubenswrapper[4899]: I0123 08:55:11.227692 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:55:12 crc kubenswrapper[4899]: I0123 08:55:12.267239 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zcjjh" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="registry-server" probeResult="failure" output=< Jan 23 08:55:12 crc kubenswrapper[4899]: timeout: failed to connect service ":50051" within 1s Jan 23 08:55:12 crc kubenswrapper[4899]: > Jan 23 08:55:13 crc kubenswrapper[4899]: I0123 08:55:13.951160 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:13 crc kubenswrapper[4899]: I0123 08:55:13.989875 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:14 crc kubenswrapper[4899]: I0123 08:55:14.813951 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerStarted","Data":"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3"} Jan 23 08:55:14 crc kubenswrapper[4899]: I0123 08:55:14.831850 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9f6g6" podStartSLOduration=3.780375216 podStartE2EDuration="54.831829329s" podCreationTimestamp="2026-01-23 08:54:20 +0000 UTC" firstStartedPulling="2026-01-23 08:54:23.329229186 +0000 UTC m=+150.196199573" lastFinishedPulling="2026-01-23 08:55:14.380683299 +0000 UTC m=+201.247653686" observedRunningTime="2026-01-23 08:55:14.831577793 +0000 UTC m=+201.698548180" watchObservedRunningTime="2026-01-23 08:55:14.831829329 +0000 UTC m=+201.698799716" Jan 23 08:55:15 crc kubenswrapper[4899]: E0123 08:55:15.142667 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:55:15 crc kubenswrapper[4899]: I0123 08:55:15.820795 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerStarted","Data":"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7"} Jan 23 08:55:15 crc kubenswrapper[4899]: I0123 08:55:15.823040 4899 generic.go:334] "Generic (PLEG): container finished" podID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerID="4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc" exitCode=0 Jan 23 08:55:15 crc kubenswrapper[4899]: I0123 08:55:15.823082 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerDied","Data":"4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc"} Jan 23 08:55:15 crc kubenswrapper[4899]: I0123 08:55:15.827847 4899 generic.go:334] "Generic (PLEG): container finished" podID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerID="775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0" exitCode=0 Jan 23 08:55:15 crc kubenswrapper[4899]: I0123 08:55:15.827894 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerDied","Data":"775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0"} Jan 23 08:55:16 crc kubenswrapper[4899]: I0123 08:55:16.812021 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:55:16 crc kubenswrapper[4899]: I0123 08:55:16.812377 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mdwxw" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="registry-server" containerID="cri-o://7bcc3c44821a82e1216c4397f634d0f9d58f889eb4772664c2c5f0a458a28dc7" gracePeriod=2 Jan 23 08:55:16 crc kubenswrapper[4899]: I0123 08:55:16.833736 4899 generic.go:334] "Generic (PLEG): container finished" podID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerID="cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7" exitCode=0 Jan 23 08:55:16 crc kubenswrapper[4899]: I0123 08:55:16.833799 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerDied","Data":"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7"} Jan 23 08:55:18 crc kubenswrapper[4899]: I0123 08:55:18.845818 4899 generic.go:334] "Generic (PLEG): container finished" podID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerID="7bcc3c44821a82e1216c4397f634d0f9d58f889eb4772664c2c5f0a458a28dc7" exitCode=0 Jan 23 08:55:18 crc kubenswrapper[4899]: I0123 08:55:18.845881 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerDied","Data":"7bcc3c44821a82e1216c4397f634d0f9d58f889eb4772664c2c5f0a458a28dc7"} Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.007861 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.138933 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q99w\" (UniqueName: \"kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w\") pod \"8e52c267-bc88-4135-a08c-90437c94c3c9\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.139160 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content\") pod \"8e52c267-bc88-4135-a08c-90437c94c3c9\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.139256 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities\") pod \"8e52c267-bc88-4135-a08c-90437c94c3c9\" (UID: \"8e52c267-bc88-4135-a08c-90437c94c3c9\") " Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.139889 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities" (OuterVolumeSpecName: "utilities") pod "8e52c267-bc88-4135-a08c-90437c94c3c9" (UID: "8e52c267-bc88-4135-a08c-90437c94c3c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.144648 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w" (OuterVolumeSpecName: "kube-api-access-7q99w") pod "8e52c267-bc88-4135-a08c-90437c94c3c9" (UID: "8e52c267-bc88-4135-a08c-90437c94c3c9"). InnerVolumeSpecName "kube-api-access-7q99w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.240453 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q99w\" (UniqueName: \"kubernetes.io/projected/8e52c267-bc88-4135-a08c-90437c94c3c9-kube-api-access-7q99w\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.240489 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.261175 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e52c267-bc88-4135-a08c-90437c94c3c9" (UID: "8e52c267-bc88-4135-a08c-90437c94c3c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.342023 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e52c267-bc88-4135-a08c-90437c94c3c9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.853894 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdwxw" event={"ID":"8e52c267-bc88-4135-a08c-90437c94c3c9","Type":"ContainerDied","Data":"5e093e40763c7cb775bb91a6ba3b41886fb2fb5f323beea87c211bee930c53ed"} Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.854632 4899 scope.go:117] "RemoveContainer" containerID="7bcc3c44821a82e1216c4397f634d0f9d58f889eb4772664c2c5f0a458a28dc7" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.853963 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdwxw" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.873514 4899 scope.go:117] "RemoveContainer" containerID="c7f0ea4ae1e11d069900075f8709c4a85c4bb939f4e27cddfb8a470044a4705f" Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.874402 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.878402 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mdwxw"] Jan 23 08:55:19 crc kubenswrapper[4899]: I0123 08:55:19.898411 4899 scope.go:117] "RemoveContainer" containerID="8f5087e1f8c7b43f0a88c063ea996d3969b9f0990b40e3e154fc9446c528189e" Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.713625 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.714018 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.774742 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.863958 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.865161 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerName="controller-manager" containerID="cri-o://275d6d681c9f4024a801d4b8bacedf350e93ccfa74f8259823ff9b47dcb12b2e" gracePeriod=30 Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.898374 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.898635 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" containerName="route-controller-manager" containerID="cri-o://845b24e80e656a0b5e8bbac827b5f90f5e12e0dc0df03ce635185e18f5ea9070" gracePeriod=30 Jan 23 08:55:20 crc kubenswrapper[4899]: I0123 08:55:20.915071 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.268644 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.309967 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.521582 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" path="/var/lib/kubelet/pods/8e52c267-bc88-4135-a08c-90437c94c3c9/volumes" Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.864725 4899 generic.go:334] "Generic (PLEG): container finished" podID="163c1af8-2bf4-4a45-86fe-39084913618d" containerID="845b24e80e656a0b5e8bbac827b5f90f5e12e0dc0df03ce635185e18f5ea9070" exitCode=0 Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.864832 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" event={"ID":"163c1af8-2bf4-4a45-86fe-39084913618d","Type":"ContainerDied","Data":"845b24e80e656a0b5e8bbac827b5f90f5e12e0dc0df03ce635185e18f5ea9070"} Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.871635 4899 generic.go:334] "Generic (PLEG): container finished" podID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerID="275d6d681c9f4024a801d4b8bacedf350e93ccfa74f8259823ff9b47dcb12b2e" exitCode=0 Jan 23 08:55:21 crc kubenswrapper[4899]: I0123 08:55:21.871800 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" event={"ID":"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd","Type":"ContainerDied","Data":"275d6d681c9f4024a801d4b8bacedf350e93ccfa74f8259823ff9b47dcb12b2e"} Jan 23 08:55:22 crc kubenswrapper[4899]: I0123 08:55:22.338030 4899 patch_prober.go:28] interesting pod/controller-manager-6d977694b9-px8jv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Jan 23 08:55:22 crc kubenswrapper[4899]: I0123 08:55:22.338108 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Jan 23 08:55:23 crc kubenswrapper[4899]: I0123 08:55:23.811376 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:55:23 crc kubenswrapper[4899]: I0123 08:55:23.881558 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9f6g6" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="registry-server" containerID="cri-o://2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3" gracePeriod=2 Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.611703 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.615706 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.647674 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.647954 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerName="controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.647971 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerName="controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.647983 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="extract-utilities" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.647990 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="extract-utilities" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.648002 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="extract-content" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648009 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="extract-content" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.648017 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" containerName="route-controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648023 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" containerName="route-controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.648037 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="registry-server" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648043 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="registry-server" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648136 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" containerName="route-controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648146 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" containerName="controller-manager" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648157 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e52c267-bc88-4135-a08c-90437c94c3c9" containerName="registry-server" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.648544 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.663564 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813026 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca\") pod \"163c1af8-2bf4-4a45-86fe-39084913618d\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813105 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca\") pod \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813130 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skpf9\" (UniqueName: \"kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9\") pod \"163c1af8-2bf4-4a45-86fe-39084913618d\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813191 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert\") pod \"163c1af8-2bf4-4a45-86fe-39084913618d\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813251 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config\") pod \"163c1af8-2bf4-4a45-86fe-39084913618d\" (UID: \"163c1af8-2bf4-4a45-86fe-39084913618d\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813300 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg5z6\" (UniqueName: \"kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6\") pod \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813320 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles\") pod \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813377 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert\") pod \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813400 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config\") pod \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\" (UID: \"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd\") " Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813587 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813614 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813650 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813681 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.813719 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49m9k\" (UniqueName: \"kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.814667 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca" (OuterVolumeSpecName: "client-ca") pod "163c1af8-2bf4-4a45-86fe-39084913618d" (UID: "163c1af8-2bf4-4a45-86fe-39084913618d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.814844 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca" (OuterVolumeSpecName: "client-ca") pod "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" (UID: "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.815402 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config" (OuterVolumeSpecName: "config") pod "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" (UID: "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.816275 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" (UID: "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.816401 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config" (OuterVolumeSpecName: "config") pod "163c1af8-2bf4-4a45-86fe-39084913618d" (UID: "163c1af8-2bf4-4a45-86fe-39084913618d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.820306 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" (UID: "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.820399 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6" (OuterVolumeSpecName: "kube-api-access-sg5z6") pod "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" (UID: "bef2ec06-fe5c-480b-8d84-67dbc30cbcfd"). InnerVolumeSpecName "kube-api-access-sg5z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.820411 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9" (OuterVolumeSpecName: "kube-api-access-skpf9") pod "163c1af8-2bf4-4a45-86fe-39084913618d" (UID: "163c1af8-2bf4-4a45-86fe-39084913618d"). InnerVolumeSpecName "kube-api-access-skpf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.820482 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "163c1af8-2bf4-4a45-86fe-39084913618d" (UID: "163c1af8-2bf4-4a45-86fe-39084913618d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.842792 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.886751 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" event={"ID":"163c1af8-2bf4-4a45-86fe-39084913618d","Type":"ContainerDied","Data":"cc12af7bac7f01902c09a1c05f1d42849f2714c073d53dd5ae46b668d6e77d50"} Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.886802 4899 scope.go:117] "RemoveContainer" containerID="845b24e80e656a0b5e8bbac827b5f90f5e12e0dc0df03ce635185e18f5ea9070" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.886867 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.888489 4899 generic.go:334] "Generic (PLEG): container finished" podID="56163971-32c1-44ab-9d10-d884891a2440" containerID="2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3" exitCode=0 Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.888539 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerDied","Data":"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3"} Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.888562 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9f6g6" event={"ID":"56163971-32c1-44ab-9d10-d884891a2440","Type":"ContainerDied","Data":"a3d567bda0f8df27a4d98123609afda546e083c5575e543aa78f67305f266f44"} Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.888602 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9f6g6" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.892435 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerStarted","Data":"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc"} Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.894629 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" event={"ID":"bef2ec06-fe5c-480b-8d84-67dbc30cbcfd","Type":"ContainerDied","Data":"1b9bc415422ebc06efcc5c4dd3debb2cfa862d04071829c1ee5d7040e05b7a83"} Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.894730 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d977694b9-px8jv" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.901652 4899 scope.go:117] "RemoveContainer" containerID="2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914191 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zgmpt" podStartSLOduration=4.696816417 podStartE2EDuration="1m5.914176225s" podCreationTimestamp="2026-01-23 08:54:19 +0000 UTC" firstStartedPulling="2026-01-23 08:54:23.310369208 +0000 UTC m=+150.177339595" lastFinishedPulling="2026-01-23 08:55:24.527728976 +0000 UTC m=+211.394699403" observedRunningTime="2026-01-23 08:55:24.911335875 +0000 UTC m=+211.778306262" watchObservedRunningTime="2026-01-23 08:55:24.914176225 +0000 UTC m=+211.781146612" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914487 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914539 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914580 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49m9k\" (UniqueName: \"kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914652 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914673 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914718 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914732 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skpf9\" (UniqueName: \"kubernetes.io/projected/163c1af8-2bf4-4a45-86fe-39084913618d-kube-api-access-skpf9\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914744 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/163c1af8-2bf4-4a45-86fe-39084913618d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914756 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914766 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg5z6\" (UniqueName: \"kubernetes.io/projected/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-kube-api-access-sg5z6\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914774 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914783 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914800 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.914809 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/163c1af8-2bf4-4a45-86fe-39084913618d-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.915759 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.915765 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.916453 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.925739 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.928692 4899 scope.go:117] "RemoveContainer" containerID="754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.938674 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.939358 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49m9k\" (UniqueName: \"kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k\") pod \"controller-manager-67dd565646-llrl5\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.944323 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h"] Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.955261 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.957959 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d977694b9-px8jv"] Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.964566 4899 scope.go:117] "RemoveContainer" containerID="68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.976659 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.977252 4899 scope.go:117] "RemoveContainer" containerID="2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.977629 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3\": container with ID starting with 2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3 not found: ID does not exist" containerID="2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.977685 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3"} err="failed to get container status \"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3\": rpc error: code = NotFound desc = could not find container \"2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3\": container with ID starting with 2e2d3e6676ed7f8459aad97998af6cfde2bd5a66ad76f1b8bf1b8f87a7a9f2d3 not found: ID does not exist" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.977711 4899 scope.go:117] "RemoveContainer" containerID="754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.978048 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022\": container with ID starting with 754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022 not found: ID does not exist" containerID="754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.978073 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022"} err="failed to get container status \"754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022\": rpc error: code = NotFound desc = could not find container \"754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022\": container with ID starting with 754cd3c78bee4edc1dac1ed287795a277b7b64692ba381229e293b9014130022 not found: ID does not exist" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.978089 4899 scope.go:117] "RemoveContainer" containerID="68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d" Jan 23 08:55:24 crc kubenswrapper[4899]: E0123 08:55:24.978570 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d\": container with ID starting with 68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d not found: ID does not exist" containerID="68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.978629 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d"} err="failed to get container status \"68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d\": rpc error: code = NotFound desc = could not find container \"68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d\": container with ID starting with 68fe04654464c979881b56d9bcead530ab3fb9992879fe583b89561255259b4d not found: ID does not exist" Jan 23 08:55:24 crc kubenswrapper[4899]: I0123 08:55:24.978649 4899 scope.go:117] "RemoveContainer" containerID="275d6d681c9f4024a801d4b8bacedf350e93ccfa74f8259823ff9b47dcb12b2e" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.015668 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content\") pod \"56163971-32c1-44ab-9d10-d884891a2440\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.042345 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78hst\" (UniqueName: \"kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst\") pod \"56163971-32c1-44ab-9d10-d884891a2440\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.042436 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities\") pod \"56163971-32c1-44ab-9d10-d884891a2440\" (UID: \"56163971-32c1-44ab-9d10-d884891a2440\") " Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.043612 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities" (OuterVolumeSpecName: "utilities") pod "56163971-32c1-44ab-9d10-d884891a2440" (UID: "56163971-32c1-44ab-9d10-d884891a2440"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.045461 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst" (OuterVolumeSpecName: "kube-api-access-78hst") pod "56163971-32c1-44ab-9d10-d884891a2440" (UID: "56163971-32c1-44ab-9d10-d884891a2440"). InnerVolumeSpecName "kube-api-access-78hst". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.081898 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56163971-32c1-44ab-9d10-d884891a2440" (UID: "56163971-32c1-44ab-9d10-d884891a2440"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.144264 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.144536 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78hst\" (UniqueName: \"kubernetes.io/projected/56163971-32c1-44ab-9d10-d884891a2440-kube-api-access-78hst\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.144547 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56163971-32c1-44ab-9d10-d884891a2440-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.243613 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.248897 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9f6g6"] Jan 23 08:55:25 crc kubenswrapper[4899]: E0123 08:55:25.293671 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.410231 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.522019 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" path="/var/lib/kubelet/pods/163c1af8-2bf4-4a45-86fe-39084913618d/volumes" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.523494 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56163971-32c1-44ab-9d10-d884891a2440" path="/var/lib/kubelet/pods/56163971-32c1-44ab-9d10-d884891a2440/volumes" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.524340 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef2ec06-fe5c-480b-8d84-67dbc30cbcfd" path="/var/lib/kubelet/pods/bef2ec06-fe5c-480b-8d84-67dbc30cbcfd/volumes" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.557159 4899 patch_prober.go:28] interesting pod/route-controller-manager-bb6f8cc6c-9js9h container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.557245 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-bb6f8cc6c-9js9h" podUID="163c1af8-2bf4-4a45-86fe-39084913618d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.901644 4899 generic.go:334] "Generic (PLEG): container finished" podID="d641b177-76eb-4bb8-bd36-238f8907701d" containerID="dcc9501875f681ae71b2b588893377a875a1bd6764c0cc357a3ae2bb193aa52c" exitCode=0 Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.901737 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerDied","Data":"dcc9501875f681ae71b2b588893377a875a1bd6764c0cc357a3ae2bb193aa52c"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.904136 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerStarted","Data":"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.908904 4899 generic.go:334] "Generic (PLEG): container finished" podID="a124473f-1d9e-40ff-99bb-5413ce650662" containerID="0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a" exitCode=0 Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.908986 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerDied","Data":"0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.911376 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" event={"ID":"cc06f349-46c1-4856-b499-56a581da1b6f","Type":"ContainerStarted","Data":"86b873daf372cff8e43586696c2bce2a525f600e6284a30296ee70fdd15f0e00"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.911409 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" event={"ID":"cc06f349-46c1-4856-b499-56a581da1b6f","Type":"ContainerStarted","Data":"a07d032b16d536d5eab12763600129d6e854e82d11fd1643796a9d6a87054559"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.911427 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.918356 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.919388 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerStarted","Data":"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d"} Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.947954 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" podStartSLOduration=5.947934235 podStartE2EDuration="5.947934235s" podCreationTimestamp="2026-01-23 08:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:25.946392997 +0000 UTC m=+212.813363394" watchObservedRunningTime="2026-01-23 08:55:25.947934235 +0000 UTC m=+212.814904622" Jan 23 08:55:25 crc kubenswrapper[4899]: I0123 08:55:25.978777 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vdpv6" podStartSLOduration=4.622370904 podStartE2EDuration="1m5.978756543s" podCreationTimestamp="2026-01-23 08:54:20 +0000 UTC" firstStartedPulling="2026-01-23 08:54:23.382368204 +0000 UTC m=+150.249338591" lastFinishedPulling="2026-01-23 08:55:24.738753843 +0000 UTC m=+211.605724230" observedRunningTime="2026-01-23 08:55:25.97658734 +0000 UTC m=+212.843557737" watchObservedRunningTime="2026-01-23 08:55:25.978756543 +0000 UTC m=+212.845726950" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.026343 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rstzc" podStartSLOduration=2.795593975 podStartE2EDuration="1m3.026322242s" podCreationTimestamp="2026-01-23 08:54:23 +0000 UTC" firstStartedPulling="2026-01-23 08:54:24.405115636 +0000 UTC m=+151.272086023" lastFinishedPulling="2026-01-23 08:55:24.635843903 +0000 UTC m=+211.502814290" observedRunningTime="2026-01-23 08:55:26.000855086 +0000 UTC m=+212.867825483" watchObservedRunningTime="2026-01-23 08:55:26.026322242 +0000 UTC m=+212.893292639" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.693787 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:26 crc kubenswrapper[4899]: E0123 08:55:26.694049 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="registry-server" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.694068 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="registry-server" Jan 23 08:55:26 crc kubenswrapper[4899]: E0123 08:55:26.694081 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="extract-utilities" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.694089 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="extract-utilities" Jan 23 08:55:26 crc kubenswrapper[4899]: E0123 08:55:26.694108 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="extract-content" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.694116 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="extract-content" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.694253 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="56163971-32c1-44ab-9d10-d884891a2440" containerName="registry-server" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.694671 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.698141 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.698411 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.698549 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.698421 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.699773 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.700294 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.705858 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.765330 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.765390 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.765593 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b97n\" (UniqueName: \"kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.765735 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.866953 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b97n\" (UniqueName: \"kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.867021 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.867068 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.867085 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.867855 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.868181 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.876118 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.883754 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b97n\" (UniqueName: \"kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n\") pod \"route-controller-manager-5f6fd4c74b-k5tcq\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.927637 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerStarted","Data":"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1"} Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.929660 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerStarted","Data":"6bc1bcb80d87c44098bf40a5ce89486ed92ce2385f0f0dcd204e2bae16366423"} Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.964486 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4ckf" podStartSLOduration=1.618309338 podStartE2EDuration="1m4.964469372s" podCreationTimestamp="2026-01-23 08:54:22 +0000 UTC" firstStartedPulling="2026-01-23 08:54:23.397328086 +0000 UTC m=+150.264298473" lastFinishedPulling="2026-01-23 08:55:26.74348812 +0000 UTC m=+213.610458507" observedRunningTime="2026-01-23 08:55:26.945888875 +0000 UTC m=+213.812859263" watchObservedRunningTime="2026-01-23 08:55:26.964469372 +0000 UTC m=+213.831439759" Jan 23 08:55:26 crc kubenswrapper[4899]: I0123 08:55:26.965056 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cj8l9" podStartSLOduration=2.716183361 podStartE2EDuration="1m4.965050286s" podCreationTimestamp="2026-01-23 08:54:22 +0000 UTC" firstStartedPulling="2026-01-23 08:54:24.400472711 +0000 UTC m=+151.267443098" lastFinishedPulling="2026-01-23 08:55:26.649339636 +0000 UTC m=+213.516310023" observedRunningTime="2026-01-23 08:55:26.964252707 +0000 UTC m=+213.831223114" watchObservedRunningTime="2026-01-23 08:55:26.965050286 +0000 UTC m=+213.832020673" Jan 23 08:55:27 crc kubenswrapper[4899]: I0123 08:55:27.018806 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:27 crc kubenswrapper[4899]: I0123 08:55:27.409445 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:27 crc kubenswrapper[4899]: W0123 08:55:27.412688 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf07765f_455a_44de_82fc_fd8edd27fbcc.slice/crio-1c0f5400f0dd1d94a5a901289fe0ccdbef8e854311c5fb67fbabe65de6903af6 WatchSource:0}: Error finding container 1c0f5400f0dd1d94a5a901289fe0ccdbef8e854311c5fb67fbabe65de6903af6: Status 404 returned error can't find the container with id 1c0f5400f0dd1d94a5a901289fe0ccdbef8e854311c5fb67fbabe65de6903af6 Jan 23 08:55:27 crc kubenswrapper[4899]: I0123 08:55:27.936456 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" event={"ID":"df07765f-455a-44de-82fc-fd8edd27fbcc","Type":"ContainerStarted","Data":"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4"} Jan 23 08:55:27 crc kubenswrapper[4899]: I0123 08:55:27.936759 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" event={"ID":"df07765f-455a-44de-82fc-fd8edd27fbcc","Type":"ContainerStarted","Data":"1c0f5400f0dd1d94a5a901289fe0ccdbef8e854311c5fb67fbabe65de6903af6"} Jan 23 08:55:27 crc kubenswrapper[4899]: I0123 08:55:27.954272 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" podStartSLOduration=7.954255651 podStartE2EDuration="7.954255651s" podCreationTimestamp="2026-01-23 08:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:27.950734854 +0000 UTC m=+214.817705261" watchObservedRunningTime="2026-01-23 08:55:27.954255651 +0000 UTC m=+214.821226048" Jan 23 08:55:28 crc kubenswrapper[4899]: I0123 08:55:28.945380 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:28 crc kubenswrapper[4899]: I0123 08:55:28.950551 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:30 crc kubenswrapper[4899]: I0123 08:55:30.288344 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:55:30 crc kubenswrapper[4899]: I0123 08:55:30.288682 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:55:30 crc kubenswrapper[4899]: I0123 08:55:30.328683 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:55:30 crc kubenswrapper[4899]: I0123 08:55:30.996877 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.229976 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.230031 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.281021 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.685566 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.685666 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.685744 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.686750 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 08:55:31 crc kubenswrapper[4899]: I0123 08:55:31.686879 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a" gracePeriod=600 Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.033364 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.439616 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.439889 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.476668 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.859590 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.860664 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.911182 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ctjz9"] Jan 23 08:55:32 crc kubenswrapper[4899]: I0123 08:55:32.936915 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.021810 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.022684 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.479544 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.479810 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.524569 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.986430 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a" exitCode=0 Jan 23 08:55:33 crc kubenswrapper[4899]: I0123 08:55:33.987180 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a"} Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.026770 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.212552 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.212774 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vdpv6" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="registry-server" containerID="cri-o://6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737" gracePeriod=2 Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.706969 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.796364 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content\") pod \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.796431 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities\") pod \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.796461 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjvpl\" (UniqueName: \"kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl\") pod \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\" (UID: \"19f0a6a2-df38-4b4a-8844-6a550e0d7eca\") " Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.797331 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities" (OuterVolumeSpecName: "utilities") pod "19f0a6a2-df38-4b4a-8844-6a550e0d7eca" (UID: "19f0a6a2-df38-4b4a-8844-6a550e0d7eca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.802395 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl" (OuterVolumeSpecName: "kube-api-access-wjvpl") pod "19f0a6a2-df38-4b4a-8844-6a550e0d7eca" (UID: "19f0a6a2-df38-4b4a-8844-6a550e0d7eca"). InnerVolumeSpecName "kube-api-access-wjvpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.846111 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19f0a6a2-df38-4b4a-8844-6a550e0d7eca" (UID: "19f0a6a2-df38-4b4a-8844-6a550e0d7eca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.897677 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.897736 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.897749 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjvpl\" (UniqueName: \"kubernetes.io/projected/19f0a6a2-df38-4b4a-8844-6a550e0d7eca-kube-api-access-wjvpl\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.993733 4899 generic.go:334] "Generic (PLEG): container finished" podID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerID="6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737" exitCode=0 Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.993828 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerDied","Data":"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737"} Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.993878 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdpv6" event={"ID":"19f0a6a2-df38-4b4a-8844-6a550e0d7eca","Type":"ContainerDied","Data":"7543d4e953b5fea41c893043f53e7cdd69405a730e9b8119a2a6effd47601e9d"} Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.993914 4899 scope.go:117] "RemoveContainer" containerID="6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.994993 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdpv6" Jan 23 08:55:34 crc kubenswrapper[4899]: I0123 08:55:34.995755 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b"} Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.007771 4899 scope.go:117] "RemoveContainer" containerID="775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.030204 4899 scope.go:117] "RemoveContainer" containerID="32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.031267 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.034785 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vdpv6"] Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.065306 4899 scope.go:117] "RemoveContainer" containerID="6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737" Jan 23 08:55:35 crc kubenswrapper[4899]: E0123 08:55:35.065801 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737\": container with ID starting with 6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737 not found: ID does not exist" containerID="6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.065838 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737"} err="failed to get container status \"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737\": rpc error: code = NotFound desc = could not find container \"6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737\": container with ID starting with 6b78d82c353080137520dfb760f829f7dffe1d94004dcc3f8eb9f4d5f8767737 not found: ID does not exist" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.065863 4899 scope.go:117] "RemoveContainer" containerID="775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0" Jan 23 08:55:35 crc kubenswrapper[4899]: E0123 08:55:35.066154 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0\": container with ID starting with 775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0 not found: ID does not exist" containerID="775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.066180 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0"} err="failed to get container status \"775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0\": rpc error: code = NotFound desc = could not find container \"775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0\": container with ID starting with 775fe93d6b102f198d1417abf0157d8579cc078da0f8d85a1e5dc8826c99aad0 not found: ID does not exist" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.066197 4899 scope.go:117] "RemoveContainer" containerID="32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0" Jan 23 08:55:35 crc kubenswrapper[4899]: E0123 08:55:35.066652 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0\": container with ID starting with 32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0 not found: ID does not exist" containerID="32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.066680 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0"} err="failed to get container status \"32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0\": rpc error: code = NotFound desc = could not find container \"32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0\": container with ID starting with 32abf6ad1d138a483e7d37c6b63983090835b55ec29eb43ac023323bb3cfeda0 not found: ID does not exist" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.211931 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.212482 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cj8l9" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="registry-server" containerID="cri-o://6bc1bcb80d87c44098bf40a5ce89486ed92ce2385f0f0dcd204e2bae16366423" gracePeriod=2 Jan 23 08:55:35 crc kubenswrapper[4899]: E0123 08:55:35.429252 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:55:35 crc kubenswrapper[4899]: I0123 08:55:35.521610 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" path="/var/lib/kubelet/pods/19f0a6a2-df38-4b4a-8844-6a550e0d7eca/volumes" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.032996 4899 generic.go:334] "Generic (PLEG): container finished" podID="d641b177-76eb-4bb8-bd36-238f8907701d" containerID="6bc1bcb80d87c44098bf40a5ce89486ed92ce2385f0f0dcd204e2bae16366423" exitCode=0 Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.033057 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerDied","Data":"6bc1bcb80d87c44098bf40a5ce89486ed92ce2385f0f0dcd204e2bae16366423"} Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.230157 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.330689 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content\") pod \"d641b177-76eb-4bb8-bd36-238f8907701d\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.330749 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x47fq\" (UniqueName: \"kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq\") pod \"d641b177-76eb-4bb8-bd36-238f8907701d\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.330950 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities\") pod \"d641b177-76eb-4bb8-bd36-238f8907701d\" (UID: \"d641b177-76eb-4bb8-bd36-238f8907701d\") " Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.331554 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities" (OuterVolumeSpecName: "utilities") pod "d641b177-76eb-4bb8-bd36-238f8907701d" (UID: "d641b177-76eb-4bb8-bd36-238f8907701d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.338577 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq" (OuterVolumeSpecName: "kube-api-access-x47fq") pod "d641b177-76eb-4bb8-bd36-238f8907701d" (UID: "d641b177-76eb-4bb8-bd36-238f8907701d"). InnerVolumeSpecName "kube-api-access-x47fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.357766 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d641b177-76eb-4bb8-bd36-238f8907701d" (UID: "d641b177-76eb-4bb8-bd36-238f8907701d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.432257 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x47fq\" (UniqueName: \"kubernetes.io/projected/d641b177-76eb-4bb8-bd36-238f8907701d-kube-api-access-x47fq\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.432291 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:36 crc kubenswrapper[4899]: I0123 08:55:36.432302 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d641b177-76eb-4bb8-bd36-238f8907701d-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.042118 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cj8l9" event={"ID":"d641b177-76eb-4bb8-bd36-238f8907701d","Type":"ContainerDied","Data":"ec3e3726fe0729601d421884bfe6674eb36067bb5e67573ecf4f870e3ed8eb81"} Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.042168 4899 scope.go:117] "RemoveContainer" containerID="6bc1bcb80d87c44098bf40a5ce89486ed92ce2385f0f0dcd204e2bae16366423" Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.042287 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cj8l9" Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.057867 4899 scope.go:117] "RemoveContainer" containerID="dcc9501875f681ae71b2b588893377a875a1bd6764c0cc357a3ae2bb193aa52c" Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.070340 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.079528 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cj8l9"] Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.099562 4899 scope.go:117] "RemoveContainer" containerID="621bc9c552313230bbe38a5a63ff4d1f57d47c07b366f51a05e83eccbd115afa" Jan 23 08:55:37 crc kubenswrapper[4899]: I0123 08:55:37.523044 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" path="/var/lib/kubelet/pods/d641b177-76eb-4bb8-bd36-238f8907701d/volumes" Jan 23 08:55:40 crc kubenswrapper[4899]: I0123 08:55:40.820068 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:40 crc kubenswrapper[4899]: I0123 08:55:40.820654 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" podUID="cc06f349-46c1-4856-b499-56a581da1b6f" containerName="controller-manager" containerID="cri-o://86b873daf372cff8e43586696c2bce2a525f600e6284a30296ee70fdd15f0e00" gracePeriod=30 Jan 23 08:55:40 crc kubenswrapper[4899]: I0123 08:55:40.925226 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:40 crc kubenswrapper[4899]: I0123 08:55:40.925537 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" podUID="df07765f-455a-44de-82fc-fd8edd27fbcc" containerName="route-controller-manager" containerID="cri-o://06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4" gracePeriod=30 Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.524335 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549077 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw"] Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549365 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="extract-utilities" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549384 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="extract-utilities" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549397 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="extract-utilities" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549404 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="extract-utilities" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549412 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="extract-content" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549420 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="extract-content" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549438 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549445 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549457 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549484 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549494 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="extract-content" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549501 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="extract-content" Jan 23 08:55:42 crc kubenswrapper[4899]: E0123 08:55:42.549509 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df07765f-455a-44de-82fc-fd8edd27fbcc" containerName="route-controller-manager" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549516 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="df07765f-455a-44de-82fc-fd8edd27fbcc" containerName="route-controller-manager" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549639 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="d641b177-76eb-4bb8-bd36-238f8907701d" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549655 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="df07765f-455a-44de-82fc-fd8edd27fbcc" containerName="route-controller-manager" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.549669 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f0a6a2-df38-4b4a-8844-6a550e0d7eca" containerName="registry-server" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.550088 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.569580 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw"] Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.611403 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert\") pod \"df07765f-455a-44de-82fc-fd8edd27fbcc\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.611475 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config\") pod \"df07765f-455a-44de-82fc-fd8edd27fbcc\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.611553 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b97n\" (UniqueName: \"kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n\") pod \"df07765f-455a-44de-82fc-fd8edd27fbcc\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.611623 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca\") pod \"df07765f-455a-44de-82fc-fd8edd27fbcc\" (UID: \"df07765f-455a-44de-82fc-fd8edd27fbcc\") " Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.612648 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config" (OuterVolumeSpecName: "config") pod "df07765f-455a-44de-82fc-fd8edd27fbcc" (UID: "df07765f-455a-44de-82fc-fd8edd27fbcc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.612674 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca" (OuterVolumeSpecName: "client-ca") pod "df07765f-455a-44de-82fc-fd8edd27fbcc" (UID: "df07765f-455a-44de-82fc-fd8edd27fbcc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.636337 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "df07765f-455a-44de-82fc-fd8edd27fbcc" (UID: "df07765f-455a-44de-82fc-fd8edd27fbcc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.642479 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n" (OuterVolumeSpecName: "kube-api-access-2b97n") pod "df07765f-455a-44de-82fc-fd8edd27fbcc" (UID: "df07765f-455a-44de-82fc-fd8edd27fbcc"). InnerVolumeSpecName "kube-api-access-2b97n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713498 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-config\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713557 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t6zm\" (UniqueName: \"kubernetes.io/projected/be491cb3-3b64-4cf1-aff5-264f32f720aa-kube-api-access-6t6zm\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713676 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be491cb3-3b64-4cf1-aff5-264f32f720aa-serving-cert\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713886 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-client-ca\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713946 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b97n\" (UniqueName: \"kubernetes.io/projected/df07765f-455a-44de-82fc-fd8edd27fbcc-kube-api-access-2b97n\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713957 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713966 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df07765f-455a-44de-82fc-fd8edd27fbcc-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.713975 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df07765f-455a-44de-82fc-fd8edd27fbcc-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.814722 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-client-ca\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.814771 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-config\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.814795 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t6zm\" (UniqueName: \"kubernetes.io/projected/be491cb3-3b64-4cf1-aff5-264f32f720aa-kube-api-access-6t6zm\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.814826 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be491cb3-3b64-4cf1-aff5-264f32f720aa-serving-cert\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.815722 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-client-ca\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.816284 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be491cb3-3b64-4cf1-aff5-264f32f720aa-config\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.822163 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be491cb3-3b64-4cf1-aff5-264f32f720aa-serving-cert\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.840834 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t6zm\" (UniqueName: \"kubernetes.io/projected/be491cb3-3b64-4cf1-aff5-264f32f720aa-kube-api-access-6t6zm\") pod \"route-controller-manager-8f65c75f-jmbqw\" (UID: \"be491cb3-3b64-4cf1-aff5-264f32f720aa\") " pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:42 crc kubenswrapper[4899]: I0123 08:55:42.874449 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.074441 4899 generic.go:334] "Generic (PLEG): container finished" podID="cc06f349-46c1-4856-b499-56a581da1b6f" containerID="86b873daf372cff8e43586696c2bce2a525f600e6284a30296ee70fdd15f0e00" exitCode=0 Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.074521 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" event={"ID":"cc06f349-46c1-4856-b499-56a581da1b6f","Type":"ContainerDied","Data":"86b873daf372cff8e43586696c2bce2a525f600e6284a30296ee70fdd15f0e00"} Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.076415 4899 generic.go:334] "Generic (PLEG): container finished" podID="df07765f-455a-44de-82fc-fd8edd27fbcc" containerID="06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4" exitCode=0 Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.076445 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" event={"ID":"df07765f-455a-44de-82fc-fd8edd27fbcc","Type":"ContainerDied","Data":"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4"} Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.076468 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" event={"ID":"df07765f-455a-44de-82fc-fd8edd27fbcc","Type":"ContainerDied","Data":"1c0f5400f0dd1d94a5a901289fe0ccdbef8e854311c5fb67fbabe65de6903af6"} Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.076489 4899 scope.go:117] "RemoveContainer" containerID="06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.076511 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.100270 4899 scope.go:117] "RemoveContainer" containerID="06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4" Jan 23 08:55:43 crc kubenswrapper[4899]: E0123 08:55:43.100673 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4\": container with ID starting with 06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4 not found: ID does not exist" containerID="06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.100716 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4"} err="failed to get container status \"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4\": rpc error: code = NotFound desc = could not find container \"06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4\": container with ID starting with 06cd3b5a36c0c1dcc113ae1ed4f5a6dc9b45b66fd27d1bbf8c35909c854c62b4 not found: ID does not exist" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.105707 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.107783 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6fd4c74b-k5tcq"] Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.162747 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.289151 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw"] Jan 23 08:55:43 crc kubenswrapper[4899]: W0123 08:55:43.293996 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe491cb3_3b64_4cf1_aff5_264f32f720aa.slice/crio-8bfb4aacb8e5c23f87d17753f8903ac71ec211a91fe36bffa141371c28a478ae WatchSource:0}: Error finding container 8bfb4aacb8e5c23f87d17753f8903ac71ec211a91fe36bffa141371c28a478ae: Status 404 returned error can't find the container with id 8bfb4aacb8e5c23f87d17753f8903ac71ec211a91fe36bffa141371c28a478ae Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.327261 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert\") pod \"cc06f349-46c1-4856-b499-56a581da1b6f\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.327319 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config\") pod \"cc06f349-46c1-4856-b499-56a581da1b6f\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.327368 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca\") pod \"cc06f349-46c1-4856-b499-56a581da1b6f\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.327401 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles\") pod \"cc06f349-46c1-4856-b499-56a581da1b6f\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.327430 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49m9k\" (UniqueName: \"kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k\") pod \"cc06f349-46c1-4856-b499-56a581da1b6f\" (UID: \"cc06f349-46c1-4856-b499-56a581da1b6f\") " Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.328269 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config" (OuterVolumeSpecName: "config") pod "cc06f349-46c1-4856-b499-56a581da1b6f" (UID: "cc06f349-46c1-4856-b499-56a581da1b6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.328973 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cc06f349-46c1-4856-b499-56a581da1b6f" (UID: "cc06f349-46c1-4856-b499-56a581da1b6f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.329688 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca" (OuterVolumeSpecName: "client-ca") pod "cc06f349-46c1-4856-b499-56a581da1b6f" (UID: "cc06f349-46c1-4856-b499-56a581da1b6f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.330906 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k" (OuterVolumeSpecName: "kube-api-access-49m9k") pod "cc06f349-46c1-4856-b499-56a581da1b6f" (UID: "cc06f349-46c1-4856-b499-56a581da1b6f"). InnerVolumeSpecName "kube-api-access-49m9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.331009 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cc06f349-46c1-4856-b499-56a581da1b6f" (UID: "cc06f349-46c1-4856-b499-56a581da1b6f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.428805 4899 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc06f349-46c1-4856-b499-56a581da1b6f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.428842 4899 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-config\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.428851 4899 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.428862 4899 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc06f349-46c1-4856-b499-56a581da1b6f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.428872 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49m9k\" (UniqueName: \"kubernetes.io/projected/cc06f349-46c1-4856-b499-56a581da1b6f-kube-api-access-49m9k\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:43 crc kubenswrapper[4899]: I0123 08:55:43.526257 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df07765f-455a-44de-82fc-fd8edd27fbcc" path="/var/lib/kubelet/pods/df07765f-455a-44de-82fc-fd8edd27fbcc/volumes" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.083590 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" event={"ID":"be491cb3-3b64-4cf1-aff5-264f32f720aa","Type":"ContainerStarted","Data":"76eaae6648538ffc6180c1e1fe2da629e687216409ae9d6a27dc8e24fb46ecba"} Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.083912 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" event={"ID":"be491cb3-3b64-4cf1-aff5-264f32f720aa","Type":"ContainerStarted","Data":"8bfb4aacb8e5c23f87d17753f8903ac71ec211a91fe36bffa141371c28a478ae"} Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.083939 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.090411 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" event={"ID":"cc06f349-46c1-4856-b499-56a581da1b6f","Type":"ContainerDied","Data":"a07d032b16d536d5eab12763600129d6e854e82d11fd1643796a9d6a87054559"} Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.090477 4899 scope.go:117] "RemoveContainer" containerID="86b873daf372cff8e43586696c2bce2a525f600e6284a30296ee70fdd15f0e00" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.090597 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67dd565646-llrl5" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.095685 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.108057 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8f65c75f-jmbqw" podStartSLOduration=4.108030225 podStartE2EDuration="4.108030225s" podCreationTimestamp="2026-01-23 08:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:44.105876062 +0000 UTC m=+230.972846509" watchObservedRunningTime="2026-01-23 08:55:44.108030225 +0000 UTC m=+230.975000642" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.144366 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.148599 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-67dd565646-llrl5"] Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.707979 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c7945658f-nhq6t"] Jan 23 08:55:44 crc kubenswrapper[4899]: E0123 08:55:44.708346 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc06f349-46c1-4856-b499-56a581da1b6f" containerName="controller-manager" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.708363 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc06f349-46c1-4856-b499-56a581da1b6f" containerName="controller-manager" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.708484 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc06f349-46c1-4856-b499-56a581da1b6f" containerName="controller-manager" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.709060 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.716097 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.716487 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.716958 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.717184 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.717547 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.719273 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.726234 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c7945658f-nhq6t"] Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.727345 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.849974 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-config\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.850056 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed3ab17-8e9b-4b86-8823-94dca1d98f24-serving-cert\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.850096 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6c4v\" (UniqueName: \"kubernetes.io/projected/bed3ab17-8e9b-4b86-8823-94dca1d98f24-kube-api-access-c6c4v\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.850119 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-proxy-ca-bundles\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.850531 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-client-ca\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.951912 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-client-ca\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.952004 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-config\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.952039 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed3ab17-8e9b-4b86-8823-94dca1d98f24-serving-cert\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.952067 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6c4v\" (UniqueName: \"kubernetes.io/projected/bed3ab17-8e9b-4b86-8823-94dca1d98f24-kube-api-access-c6c4v\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.952094 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-proxy-ca-bundles\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.953296 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-client-ca\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.953386 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-proxy-ca-bundles\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.953446 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed3ab17-8e9b-4b86-8823-94dca1d98f24-config\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.960515 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed3ab17-8e9b-4b86-8823-94dca1d98f24-serving-cert\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:44 crc kubenswrapper[4899]: I0123 08:55:44.972730 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6c4v\" (UniqueName: \"kubernetes.io/projected/bed3ab17-8e9b-4b86-8823-94dca1d98f24-kube-api-access-c6c4v\") pod \"controller-manager-c7945658f-nhq6t\" (UID: \"bed3ab17-8e9b-4b86-8823-94dca1d98f24\") " pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:45 crc kubenswrapper[4899]: I0123 08:55:45.033388 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:45 crc kubenswrapper[4899]: I0123 08:55:45.483878 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c7945658f-nhq6t"] Jan 23 08:55:45 crc kubenswrapper[4899]: I0123 08:55:45.525853 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc06f349-46c1-4856-b499-56a581da1b6f" path="/var/lib/kubelet/pods/cc06f349-46c1-4856-b499-56a581da1b6f/volumes" Jan 23 08:55:45 crc kubenswrapper[4899]: E0123 08:55:45.559933 4899 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a08d77e_7f9c_4c41_8d85_95b64e461ea5.slice/crio-6637997b369bee6f864c5c124762253073d24e59c99b1a6266baefaaaf82fd53\": RecentStats: unable to find data in memory cache]" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.108264 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" event={"ID":"bed3ab17-8e9b-4b86-8823-94dca1d98f24","Type":"ContainerStarted","Data":"54de98435a2acb0d77ce5818d6f73febb3459df21f9c95aa902f4b09b30a3765"} Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.108761 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.108793 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" event={"ID":"bed3ab17-8e9b-4b86-8823-94dca1d98f24","Type":"ContainerStarted","Data":"9cfcd4c511fc5c3d03dd3f2bc350cfa3dcc3f3995766f03bcf4ab85998296fbc"} Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.114083 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.127149 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c7945658f-nhq6t" podStartSLOduration=6.127130284 podStartE2EDuration="6.127130284s" podCreationTimestamp="2026-01-23 08:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:55:46.124673224 +0000 UTC m=+232.991643611" watchObservedRunningTime="2026-01-23 08:55:46.127130284 +0000 UTC m=+232.994100671" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.720490 4899 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721435 4899 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721633 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721819 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d" gracePeriod=15 Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721824 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012" gracePeriod=15 Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721957 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db" gracePeriod=15 Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.721998 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd" gracePeriod=15 Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.722018 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648" gracePeriod=15 Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725011 4899 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725369 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725388 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725419 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725431 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725442 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725451 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725460 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725469 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725478 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725488 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725512 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725519 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725529 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725538 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725703 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725732 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725745 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725754 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725767 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725776 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 23 08:55:46 crc kubenswrapper[4899]: E0123 08:55:46.725913 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.725927 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.726071 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.768890 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876362 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876422 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876446 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876483 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876498 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876520 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876546 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.876562 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978128 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978188 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978229 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978258 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978277 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978376 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978412 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978370 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978316 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978391 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978490 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978513 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978346 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978573 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978440 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:46 crc kubenswrapper[4899]: I0123 08:55:46.978578 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.062387 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:55:47 crc kubenswrapper[4899]: W0123 08:55:47.090897 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-dc4ff19d458d5b765f90db40e8ca5aa122fca17921b6d8b4323d061dfa92dbda WatchSource:0}: Error finding container dc4ff19d458d5b765f90db40e8ca5aa122fca17921b6d8b4323d061dfa92dbda: Status 404 returned error can't find the container with id dc4ff19d458d5b765f90db40e8ca5aa122fca17921b6d8b4323d061dfa92dbda Jan 23 08:55:47 crc kubenswrapper[4899]: E0123 08:55:47.094615 4899 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d50533edd390e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-23 08:55:47.093666062 +0000 UTC m=+233.960636479,LastTimestamp:2026-01-23 08:55:47.093666062 +0000 UTC m=+233.960636479,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.117392 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.119161 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.120387 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012" exitCode=0 Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.120410 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648" exitCode=0 Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.120419 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db" exitCode=0 Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.120428 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd" exitCode=2 Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.120495 4899 scope.go:117] "RemoveContainer" containerID="3cdc1d7c6e8d39a0b7c9c402406a77640d80f9ffc332443efbd7c6b3c17a0548" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.121689 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"dc4ff19d458d5b765f90db40e8ca5aa122fca17921b6d8b4323d061dfa92dbda"} Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.123434 4899 generic.go:334] "Generic (PLEG): container finished" podID="68143dd5-8349-40d1-ac14-9d30e14ad3af" containerID="ad7fa93087cbb5c11593435e5a9e42f0f67be3e409764682f63ef2401be0b899" exitCode=0 Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.123482 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"68143dd5-8349-40d1-ac14-9d30e14ad3af","Type":"ContainerDied","Data":"ad7fa93087cbb5c11593435e5a9e42f0f67be3e409764682f63ef2401be0b899"} Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.124320 4899 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.124844 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:47 crc kubenswrapper[4899]: I0123 08:55:47.125112 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.132090 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2"} Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.133076 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.133701 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.137033 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.462118 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.463496 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.463733 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.606819 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access\") pod \"68143dd5-8349-40d1-ac14-9d30e14ad3af\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.606988 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock\") pod \"68143dd5-8349-40d1-ac14-9d30e14ad3af\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.607012 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir\") pod \"68143dd5-8349-40d1-ac14-9d30e14ad3af\" (UID: \"68143dd5-8349-40d1-ac14-9d30e14ad3af\") " Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.607070 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock" (OuterVolumeSpecName: "var-lock") pod "68143dd5-8349-40d1-ac14-9d30e14ad3af" (UID: "68143dd5-8349-40d1-ac14-9d30e14ad3af"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.607152 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "68143dd5-8349-40d1-ac14-9d30e14ad3af" (UID: "68143dd5-8349-40d1-ac14-9d30e14ad3af"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.607362 4899 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-var-lock\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.607378 4899 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68143dd5-8349-40d1-ac14-9d30e14ad3af-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.613775 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "68143dd5-8349-40d1-ac14-9d30e14ad3af" (UID: "68143dd5-8349-40d1-ac14-9d30e14ad3af"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:48 crc kubenswrapper[4899]: I0123 08:55:48.708788 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68143dd5-8349-40d1-ac14-9d30e14ad3af-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.083531 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.084664 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.085187 4899 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.085455 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.085662 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.145428 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"68143dd5-8349-40d1-ac14-9d30e14ad3af","Type":"ContainerDied","Data":"bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe"} Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.145492 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb24ae333aa146f5f69de1c8fadaad981a34e2500a74715d8867fd2a87f68bfe" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.145444 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.148859 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.149716 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d" exitCode=0 Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.149805 4899 scope.go:117] "RemoveContainer" containerID="5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.149902 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.162516 4899 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.162941 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.163193 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.167498 4899 scope.go:117] "RemoveContainer" containerID="bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.181293 4899 scope.go:117] "RemoveContainer" containerID="66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.194849 4899 scope.go:117] "RemoveContainer" containerID="ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.209669 4899 scope.go:117] "RemoveContainer" containerID="1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.232857 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.232886 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233003 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233029 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233044 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233148 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233285 4899 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233301 4899 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.233314 4899 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.234564 4899 scope.go:117] "RemoveContainer" containerID="a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.260505 4899 scope.go:117] "RemoveContainer" containerID="5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.264326 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\": container with ID starting with 5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012 not found: ID does not exist" containerID="5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.264399 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012"} err="failed to get container status \"5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\": rpc error: code = NotFound desc = could not find container \"5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012\": container with ID starting with 5d9bda5950da8afa490cc7a609ab1fc162266df6c9fa8538119f9ae5effb3012 not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.264438 4899 scope.go:117] "RemoveContainer" containerID="bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.264791 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\": container with ID starting with bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648 not found: ID does not exist" containerID="bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.264830 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648"} err="failed to get container status \"bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\": rpc error: code = NotFound desc = could not find container \"bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648\": container with ID starting with bee3681f4bc57c654c7b7b5a9c1184dedca801bd436910de1120f43503a83648 not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.264858 4899 scope.go:117] "RemoveContainer" containerID="66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.265286 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\": container with ID starting with 66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db not found: ID does not exist" containerID="66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.265318 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db"} err="failed to get container status \"66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\": rpc error: code = NotFound desc = could not find container \"66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db\": container with ID starting with 66b34d95c3d2664cd383502ce09430a736513d779daac8705c8517b8c05098db not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.265339 4899 scope.go:117] "RemoveContainer" containerID="ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.265652 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\": container with ID starting with ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd not found: ID does not exist" containerID="ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.265683 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd"} err="failed to get container status \"ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\": rpc error: code = NotFound desc = could not find container \"ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd\": container with ID starting with ea5e974ceb99a06fb6003ffc79d479fe6945569014f0872a5a64475109ed2dfd not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.265703 4899 scope.go:117] "RemoveContainer" containerID="1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.266030 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\": container with ID starting with 1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d not found: ID does not exist" containerID="1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.266073 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d"} err="failed to get container status \"1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\": rpc error: code = NotFound desc = could not find container \"1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d\": container with ID starting with 1322759c124cf030ac3bf4d9a178ecf5a85750c78c8cc20914fb08e40bb7c00d not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.266097 4899 scope.go:117] "RemoveContainer" containerID="a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834" Jan 23 08:55:49 crc kubenswrapper[4899]: E0123 08:55:49.266416 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\": container with ID starting with a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834 not found: ID does not exist" containerID="a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.266442 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834"} err="failed to get container status \"a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\": rpc error: code = NotFound desc = could not find container \"a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834\": container with ID starting with a4d77283a39a95ee9b8027b54f72d593444676128261e56beecb1132c7dd8834 not found: ID does not exist" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.478346 4899 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.478913 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.479494 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:49 crc kubenswrapper[4899]: I0123 08:55:49.527228 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 23 08:55:51 crc kubenswrapper[4899]: E0123 08:55:51.812584 4899 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d50533edd390e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-23 08:55:47.093666062 +0000 UTC m=+233.960636479,LastTimestamp:2026-01-23 08:55:47.093666062 +0000 UTC m=+233.960636479,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 23 08:55:53 crc kubenswrapper[4899]: I0123 08:55:53.522000 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:53 crc kubenswrapper[4899]: I0123 08:55:53.522630 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.557854 4899 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.558140 4899 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.558422 4899 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.558670 4899 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.558906 4899 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:56 crc kubenswrapper[4899]: I0123 08:55:56.558949 4899 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.559193 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="200ms" Jan 23 08:55:56 crc kubenswrapper[4899]: E0123 08:55:56.760897 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="400ms" Jan 23 08:55:57 crc kubenswrapper[4899]: E0123 08:55:57.162448 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="800ms" Jan 23 08:55:57 crc kubenswrapper[4899]: I0123 08:55:57.945165 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" containerID="cri-o://a46826f41d8c9d5ae450062b940e8aa27584cc5bf990d48c3ce55b7f7a6e284f" gracePeriod=15 Jan 23 08:55:57 crc kubenswrapper[4899]: E0123 08:55:57.964043 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="1.6s" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.203805 4899 generic.go:334] "Generic (PLEG): container finished" podID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerID="a46826f41d8c9d5ae450062b940e8aa27584cc5bf990d48c3ce55b7f7a6e284f" exitCode=0 Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.203850 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" event={"ID":"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0","Type":"ContainerDied","Data":"a46826f41d8c9d5ae450062b940e8aa27584cc5bf990d48c3ce55b7f7a6e284f"} Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.530293 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.531247 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.531534 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.531791 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658500 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658568 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658607 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658645 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnp29\" (UniqueName: \"kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658679 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658704 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658746 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658773 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658798 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658819 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658844 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658870 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658905 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658944 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection\") pod \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\" (UID: \"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0\") " Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.658935 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.659785 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.659926 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.660573 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.661129 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.665917 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.666183 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29" (OuterVolumeSpecName: "kube-api-access-tnp29") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "kube-api-access-tnp29". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.666234 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.666398 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.672646 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.673140 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.673566 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.673709 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.674448 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" (UID: "1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761111 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761148 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761158 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761167 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnp29\" (UniqueName: \"kubernetes.io/projected/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-kube-api-access-tnp29\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761176 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761187 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761198 4899 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761246 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761256 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761266 4899 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761274 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761283 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761292 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:58 crc kubenswrapper[4899]: I0123 08:55:58.761301 4899 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.211853 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" event={"ID":"1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0","Type":"ContainerDied","Data":"fafbdb3a2bc3bbaae2b6592ce1f1fe7c56e3ad488089e09467f1f603d46168b3"} Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.211907 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.211963 4899 scope.go:117] "RemoveContainer" containerID="a46826f41d8c9d5ae450062b940e8aa27584cc5bf990d48c3ce55b7f7a6e284f" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.214121 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.214772 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.215312 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.225993 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.226247 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: I0123 08:55:59.226462 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:55:59 crc kubenswrapper[4899]: E0123 08:55:59.565254 4899 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="3.2s" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.513903 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.516320 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.516968 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.517873 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.543988 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.544032 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:00 crc kubenswrapper[4899]: E0123 08:56:00.544498 4899 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:00 crc kubenswrapper[4899]: I0123 08:56:00.545106 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:00 crc kubenswrapper[4899]: E0123 08:56:00.591693 4899 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" volumeName="registry-storage" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.228550 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.228844 4899 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90" exitCode=1 Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.228975 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90"} Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.229606 4899 scope.go:117] "RemoveContainer" containerID="0f61037ed8fed5dd3753cdbcedaf544775f528e4b7ac4b88533867842e540c90" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.229834 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.230360 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.230972 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.231478 4899 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.232478 4899 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ecf66242c3ec058573571e530f9f3d76fdd0bc2f878b6b53f79626155bf2c883" exitCode=0 Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.232525 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ecf66242c3ec058573571e530f9f3d76fdd0bc2f878b6b53f79626155bf2c883"} Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.232559 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c61fc4d5ef67520b3479c7d6968ba23b564e3fe4c45e19ddfe1a5b229f94f3b6"} Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.232812 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.232836 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:01 crc kubenswrapper[4899]: E0123 08:56:01.233106 4899 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.233229 4899 status_manager.go:851] "Failed to get status for pod" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" pod="openshift-authentication/oauth-openshift-558db77b4-ctjz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ctjz9\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.233658 4899 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.234004 4899 status_manager.go:851] "Failed to get status for pod" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:01 crc kubenswrapper[4899]: I0123 08:56:01.234385 4899 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Jan 23 08:56:02 crc kubenswrapper[4899]: I0123 08:56:02.118372 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:56:02 crc kubenswrapper[4899]: I0123 08:56:02.238969 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 23 08:56:02 crc kubenswrapper[4899]: I0123 08:56:02.239074 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"31a343673d93cf4d539cb22206ca5fbd2d9eb5e71b0d0b17eba4bba39c7b16e9"} Jan 23 08:56:02 crc kubenswrapper[4899]: I0123 08:56:02.241814 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7666d5bfd74e45c8897c62bc4f68188bd3583ccefe4e82dafae1e79566135001"} Jan 23 08:56:02 crc kubenswrapper[4899]: I0123 08:56:02.241847 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ee3dc1f423a82b7cb07ff3800e04e57c85b445e60fefcae34be6713c1ce58d8d"} Jan 23 08:56:03 crc kubenswrapper[4899]: I0123 08:56:03.256798 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2fbcd1dd58eafb2f068bfd11cf94e04a1e28f68b33fb73f9ae15c98dc22ae8af"} Jan 23 08:56:03 crc kubenswrapper[4899]: I0123 08:56:03.257155 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3dcd1568ae39fed60ba4827d91e9078aafe2d6df45a11ef68dec1c1b9c5880c2"} Jan 23 08:56:03 crc kubenswrapper[4899]: I0123 08:56:03.257172 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"eb8c44ac9ff7bd9b65e0ad934305b713bf4bd69e37e76d85960ced5312fcee40"} Jan 23 08:56:03 crc kubenswrapper[4899]: I0123 08:56:03.257195 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:03 crc kubenswrapper[4899]: I0123 08:56:03.257247 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:05 crc kubenswrapper[4899]: I0123 08:56:05.546329 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:05 crc kubenswrapper[4899]: I0123 08:56:05.546828 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:05 crc kubenswrapper[4899]: I0123 08:56:05.552883 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:07 crc kubenswrapper[4899]: I0123 08:56:07.323036 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:56:07 crc kubenswrapper[4899]: I0123 08:56:07.327503 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.266043 4899 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.284055 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.284101 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.284113 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.284147 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.287677 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:08 crc kubenswrapper[4899]: I0123 08:56:08.291351 4899 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7c5aa455-3aa4-496a-bfb4-12a9546cb1b4" Jan 23 08:56:09 crc kubenswrapper[4899]: I0123 08:56:09.288002 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:09 crc kubenswrapper[4899]: I0123 08:56:09.288032 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:12 crc kubenswrapper[4899]: I0123 08:56:12.122778 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 23 08:56:13 crc kubenswrapper[4899]: I0123 08:56:13.539179 4899 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7c5aa455-3aa4-496a-bfb4-12a9546cb1b4" Jan 23 08:56:17 crc kubenswrapper[4899]: I0123 08:56:17.809308 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 23 08:56:18 crc kubenswrapper[4899]: I0123 08:56:18.316335 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 23 08:56:18 crc kubenswrapper[4899]: I0123 08:56:18.370717 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 23 08:56:18 crc kubenswrapper[4899]: I0123 08:56:18.608042 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 23 08:56:18 crc kubenswrapper[4899]: I0123 08:56:18.750663 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 23 08:56:19 crc kubenswrapper[4899]: I0123 08:56:19.331836 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 23 08:56:19 crc kubenswrapper[4899]: I0123 08:56:19.439930 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.057893 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.068621 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.103638 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.274682 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.278255 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.305670 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.330659 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.333347 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.694479 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.871849 4899 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.961620 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 23 08:56:20 crc kubenswrapper[4899]: I0123 08:56:20.972658 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.287767 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.337254 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.338450 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.411492 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.436282 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.520644 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.637448 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.726613 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.819578 4899 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 23 08:56:21 crc kubenswrapper[4899]: I0123 08:56:21.929573 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.039378 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.178187 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.209040 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.262764 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.274648 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.331987 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.469622 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.572417 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.580318 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.647820 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.735120 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.781911 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.837322 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.851456 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.860846 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.902045 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.916053 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.939898 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.950351 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 23 08:56:22 crc kubenswrapper[4899]: I0123 08:56:22.976953 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.003290 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.089342 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.230990 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.284627 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.304089 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.313220 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.350118 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.474423 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.501003 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.549537 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.639036 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.663296 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.728903 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.745120 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.804197 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.887145 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.895118 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.952505 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 23 08:56:23 crc kubenswrapper[4899]: I0123 08:56:23.956159 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.034966 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.039996 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.045830 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.054846 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.173573 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.306474 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.337691 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.358367 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.381565 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.415261 4899 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.431165 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.431137485 podStartE2EDuration="38.431137485s" podCreationTimestamp="2026-01-23 08:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:56:08.148469843 +0000 UTC m=+255.015440240" watchObservedRunningTime="2026-01-23 08:56:24.431137485 +0000 UTC m=+271.298107882" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432314 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ctjz9","openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432383 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-85d79b69d6-jjntt","openshift-kube-apiserver/kube-apiserver-crc"] Jan 23 08:56:24 crc kubenswrapper[4899]: E0123 08:56:24.432662 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432683 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" Jan 23 08:56:24 crc kubenswrapper[4899]: E0123 08:56:24.432702 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" containerName="installer" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432711 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" containerName="installer" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432836 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="68143dd5-8349-40d1-ac14-9d30e14ad3af" containerName="installer" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.432859 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" containerName="oauth-openshift" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.433109 4899 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.433170 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d29358f2-5909-45b8-bacf-e12b71fea05d" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.433578 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.437871 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.437972 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.438092 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.438172 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.438647 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.438838 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.438901 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.439042 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.439383 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.439850 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.440360 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.441499 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.442791 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.451198 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.454409 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.457508 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.477862 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.492540 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.499834 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.49979704 podStartE2EDuration="16.49979704s" podCreationTimestamp="2026-01-23 08:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:56:24.481520056 +0000 UTC m=+271.348490443" watchObservedRunningTime="2026-01-23 08:56:24.49979704 +0000 UTC m=+271.366767467" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.529442 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.604713 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-dir\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.604944 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.604978 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.604997 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605024 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605045 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605434 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-policies\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605659 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605694 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605716 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-error\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605751 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grx9b\" (UniqueName: \"kubernetes.io/projected/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-kube-api-access-grx9b\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605781 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605800 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-session\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.605827 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-login\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.656648 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.662398 4899 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.697137 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706552 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706619 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-error\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706648 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grx9b\" (UniqueName: \"kubernetes.io/projected/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-kube-api-access-grx9b\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706680 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706701 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-session\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706729 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-login\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706751 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-dir\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706784 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706810 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706832 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706864 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.706889 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.707317 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-dir\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.707913 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-policies\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.707945 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.708298 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.708719 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.708821 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.709075 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-audit-policies\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714077 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-error\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714110 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714342 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714361 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-session\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.714469 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-user-template-login\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.715247 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.720628 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.730707 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grx9b\" (UniqueName: \"kubernetes.io/projected/012e46d5-1a3c-42cc-bb74-cc30d8e4385f-kube-api-access-grx9b\") pod \"oauth-openshift-85d79b69d6-jjntt\" (UID: \"012e46d5-1a3c-42cc-bb74-cc30d8e4385f\") " pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.756053 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.804089 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.886030 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 23 08:56:24 crc kubenswrapper[4899]: I0123 08:56:24.915001 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.088542 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.090928 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.125570 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.163770 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.300486 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.305383 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.332863 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.403507 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.407436 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.437598 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.522126 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0" path="/var/lib/kubelet/pods/1258f97e-0f29-4f1d-8cf0-ceeb0a7c5ea0/volumes" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.566508 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.668408 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.763698 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.772019 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.829177 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.904856 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.909445 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.976912 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 23 08:56:25 crc kubenswrapper[4899]: I0123 08:56:25.988998 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.002907 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.014858 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.049518 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.050110 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.056181 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.097268 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.112498 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.116562 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.129570 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.358347 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.368566 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.397057 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.400983 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.439816 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.448822 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.513043 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.536937 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.560634 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.593708 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.710573 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.736616 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.760660 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.794989 4899 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 23 08:56:26 crc kubenswrapper[4899]: I0123 08:56:26.844939 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.065852 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.090500 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.091738 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.142244 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.191261 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.204847 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.214024 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.270400 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.270960 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.373803 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.421078 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.477483 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.568532 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.657360 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.681534 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.761732 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.906684 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 23 08:56:27 crc kubenswrapper[4899]: I0123 08:56:27.945460 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.055983 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.082726 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.109935 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.110157 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.163484 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.164078 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.173443 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 23 08:56:28 crc kubenswrapper[4899]: E0123 08:56:28.189355 4899 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 23 08:56:28 crc kubenswrapper[4899]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-85d79b69d6-jjntt_openshift-authentication_012e46d5-1a3c-42cc-bb74-cc30d8e4385f_0(6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5): error adding pod openshift-authentication_oauth-openshift-85d79b69d6-jjntt to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5" Netns:"/var/run/netns/4fce9428-852d-48ef-a6a9-c526bfaed56e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-85d79b69d6-jjntt;K8S_POD_INFRA_CONTAINER_ID=6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5;K8S_POD_UID=012e46d5-1a3c-42cc-bb74-cc30d8e4385f" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-85d79b69d6-jjntt] networking: Multus: [openshift-authentication/oauth-openshift-85d79b69d6-jjntt/012e46d5-1a3c-42cc-bb74-cc30d8e4385f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-85d79b69d6-jjntt in out of cluster comm: pod "oauth-openshift-85d79b69d6-jjntt" not found Jan 23 08:56:28 crc kubenswrapper[4899]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 23 08:56:28 crc kubenswrapper[4899]: > Jan 23 08:56:28 crc kubenswrapper[4899]: E0123 08:56:28.189438 4899 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 23 08:56:28 crc kubenswrapper[4899]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-85d79b69d6-jjntt_openshift-authentication_012e46d5-1a3c-42cc-bb74-cc30d8e4385f_0(6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5): error adding pod openshift-authentication_oauth-openshift-85d79b69d6-jjntt to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5" Netns:"/var/run/netns/4fce9428-852d-48ef-a6a9-c526bfaed56e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-85d79b69d6-jjntt;K8S_POD_INFRA_CONTAINER_ID=6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5;K8S_POD_UID=012e46d5-1a3c-42cc-bb74-cc30d8e4385f" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-85d79b69d6-jjntt] networking: Multus: [openshift-authentication/oauth-openshift-85d79b69d6-jjntt/012e46d5-1a3c-42cc-bb74-cc30d8e4385f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-85d79b69d6-jjntt in out of cluster comm: pod "oauth-openshift-85d79b69d6-jjntt" not found Jan 23 08:56:28 crc kubenswrapper[4899]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 23 08:56:28 crc kubenswrapper[4899]: > pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:28 crc kubenswrapper[4899]: E0123 08:56:28.189459 4899 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 23 08:56:28 crc kubenswrapper[4899]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-85d79b69d6-jjntt_openshift-authentication_012e46d5-1a3c-42cc-bb74-cc30d8e4385f_0(6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5): error adding pod openshift-authentication_oauth-openshift-85d79b69d6-jjntt to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5" Netns:"/var/run/netns/4fce9428-852d-48ef-a6a9-c526bfaed56e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-85d79b69d6-jjntt;K8S_POD_INFRA_CONTAINER_ID=6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5;K8S_POD_UID=012e46d5-1a3c-42cc-bb74-cc30d8e4385f" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-85d79b69d6-jjntt] networking: Multus: [openshift-authentication/oauth-openshift-85d79b69d6-jjntt/012e46d5-1a3c-42cc-bb74-cc30d8e4385f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-85d79b69d6-jjntt in out of cluster comm: pod "oauth-openshift-85d79b69d6-jjntt" not found Jan 23 08:56:28 crc kubenswrapper[4899]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 23 08:56:28 crc kubenswrapper[4899]: > pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:28 crc kubenswrapper[4899]: E0123 08:56:28.189521 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-85d79b69d6-jjntt_openshift-authentication(012e46d5-1a3c-42cc-bb74-cc30d8e4385f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-85d79b69d6-jjntt_openshift-authentication(012e46d5-1a3c-42cc-bb74-cc30d8e4385f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-85d79b69d6-jjntt_openshift-authentication_012e46d5-1a3c-42cc-bb74-cc30d8e4385f_0(6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5): error adding pod openshift-authentication_oauth-openshift-85d79b69d6-jjntt to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5\\\" Netns:\\\"/var/run/netns/4fce9428-852d-48ef-a6a9-c526bfaed56e\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-85d79b69d6-jjntt;K8S_POD_INFRA_CONTAINER_ID=6739a53ebc66fa0dc06675e51c8f1a51c1c1d1338abbe2012729ae8b74143ea5;K8S_POD_UID=012e46d5-1a3c-42cc-bb74-cc30d8e4385f\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-85d79b69d6-jjntt] networking: Multus: [openshift-authentication/oauth-openshift-85d79b69d6-jjntt/012e46d5-1a3c-42cc-bb74-cc30d8e4385f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-85d79b69d6-jjntt in out of cluster comm: pod \\\"oauth-openshift-85d79b69d6-jjntt\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" podUID="012e46d5-1a3c-42cc-bb74-cc30d8e4385f" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.409683 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.610926 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.736715 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.882295 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.967264 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.983274 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 23 08:56:28 crc kubenswrapper[4899]: I0123 08:56:28.988388 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.091099 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.231955 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.271533 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.291543 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.330155 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.334741 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.374977 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.377644 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.445592 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.496549 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.546918 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.592839 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.738646 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.742721 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.779179 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.841954 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.894577 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.917491 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.959376 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.963644 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.971201 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.977384 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 23 08:56:29 crc kubenswrapper[4899]: I0123 08:56:29.986569 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.010512 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.030152 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.146574 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.180111 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.250921 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.257317 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.358501 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.423900 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.583386 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.733362 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.735777 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.747828 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.802891 4899 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.803254 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2" gracePeriod=5 Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.812189 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 23 08:56:30 crc kubenswrapper[4899]: I0123 08:56:30.964337 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.033332 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.049073 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.059400 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.138771 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.232056 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.248309 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.274104 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.368800 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.398654 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.509047 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.596160 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.735046 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.749163 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.752009 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.771952 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.782127 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.788549 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.822870 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.957932 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 23 08:56:31 crc kubenswrapper[4899]: I0123 08:56:31.967019 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.192525 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.256362 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.307571 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.429156 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.525593 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.625004 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.676771 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.802342 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 23 08:56:32 crc kubenswrapper[4899]: I0123 08:56:32.810049 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 23 08:56:33 crc kubenswrapper[4899]: I0123 08:56:33.262539 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 23 08:56:33 crc kubenswrapper[4899]: I0123 08:56:33.793479 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 23 08:56:33 crc kubenswrapper[4899]: I0123 08:56:33.936139 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 23 08:56:33 crc kubenswrapper[4899]: I0123 08:56:33.963567 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 23 08:56:34 crc kubenswrapper[4899]: I0123 08:56:34.235730 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 23 08:56:34 crc kubenswrapper[4899]: I0123 08:56:34.469619 4899 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 23 08:56:34 crc kubenswrapper[4899]: I0123 08:56:34.688104 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 23 08:56:35 crc kubenswrapper[4899]: I0123 08:56:35.114435 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 23 08:56:35 crc kubenswrapper[4899]: I0123 08:56:35.216380 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.378685 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.379323 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.445311 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.445383 4899 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2" exitCode=137 Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.445441 4899 scope.go:117] "RemoveContainer" containerID="3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.445474 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.465748 4899 scope.go:117] "RemoveContainer" containerID="3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2" Jan 23 08:56:36 crc kubenswrapper[4899]: E0123 08:56:36.466716 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2\": container with ID starting with 3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2 not found: ID does not exist" containerID="3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.466756 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2"} err="failed to get container status \"3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2\": rpc error: code = NotFound desc = could not find container \"3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2\": container with ID starting with 3ddef05483aa4dac610e02749a1a3698c91fc589a11babeb29ab37570f4bb3b2 not found: ID does not exist" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557346 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557404 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557443 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557424 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557483 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557545 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557593 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557726 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.557751 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.559542 4899 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.559565 4899 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.559574 4899 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.559582 4899 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.564079 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 08:56:36 crc kubenswrapper[4899]: I0123 08:56:36.661190 4899 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.521578 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.522204 4899 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.533168 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.533230 4899 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8a83bf7e-25c8-4b82-8c30-101bb404b69c" Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.552080 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 23 08:56:37 crc kubenswrapper[4899]: I0123 08:56:37.552153 4899 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8a83bf7e-25c8-4b82-8c30-101bb404b69c" Jan 23 08:56:42 crc kubenswrapper[4899]: I0123 08:56:42.513694 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:42 crc kubenswrapper[4899]: I0123 08:56:42.514519 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:42 crc kubenswrapper[4899]: I0123 08:56:42.947242 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85d79b69d6-jjntt"] Jan 23 08:56:43 crc kubenswrapper[4899]: I0123 08:56:43.483475 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" event={"ID":"012e46d5-1a3c-42cc-bb74-cc30d8e4385f","Type":"ContainerStarted","Data":"4afc64a1ed22c52be65fd05410d5637af3cfb649cca6c4d81ae5bdc9a70e86d8"} Jan 23 08:56:43 crc kubenswrapper[4899]: I0123 08:56:43.483523 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" event={"ID":"012e46d5-1a3c-42cc-bb74-cc30d8e4385f","Type":"ContainerStarted","Data":"e4aefd1f13c06c6757f79969c5a0c9d61e179d6b1f1bf17c768b9a689dd290a4"} Jan 23 08:56:43 crc kubenswrapper[4899]: I0123 08:56:43.483951 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:43 crc kubenswrapper[4899]: I0123 08:56:43.519247 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" podStartSLOduration=71.519186302 podStartE2EDuration="1m11.519186302s" podCreationTimestamp="2026-01-23 08:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:56:43.508889736 +0000 UTC m=+290.375860153" watchObservedRunningTime="2026-01-23 08:56:43.519186302 +0000 UTC m=+290.386156779" Jan 23 08:56:43 crc kubenswrapper[4899]: I0123 08:56:43.709449 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-85d79b69d6-jjntt" Jan 23 08:56:48 crc kubenswrapper[4899]: I0123 08:56:48.514204 4899 generic.go:334] "Generic (PLEG): container finished" podID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerID="d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd" exitCode=0 Jan 23 08:56:48 crc kubenswrapper[4899]: I0123 08:56:48.514296 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerDied","Data":"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd"} Jan 23 08:56:48 crc kubenswrapper[4899]: I0123 08:56:48.515008 4899 scope.go:117] "RemoveContainer" containerID="d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd" Jan 23 08:56:49 crc kubenswrapper[4899]: I0123 08:56:49.523948 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerStarted","Data":"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662"} Jan 23 08:56:49 crc kubenswrapper[4899]: I0123 08:56:49.524709 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:56:49 crc kubenswrapper[4899]: I0123 08:56:49.529585 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:56:53 crc kubenswrapper[4899]: I0123 08:56:53.393099 4899 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.567817 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hl6gw"] Jan 23 08:57:04 crc kubenswrapper[4899]: E0123 08:57:04.568652 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.568670 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.568791 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.569249 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.579071 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hl6gw"] Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729296 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-tls\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729355 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-certificates\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729391 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729419 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-trusted-ca\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729443 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/80ff6eb9-fb27-4731-bd91-b1b73a602287-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729463 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/80ff6eb9-fb27-4731-bd91-b1b73a602287-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729478 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-bound-sa-token\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.729530 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnmp5\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-kube-api-access-xnmp5\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.760594 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830172 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-certificates\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830324 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-trusted-ca\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830361 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/80ff6eb9-fb27-4731-bd91-b1b73a602287-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830390 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/80ff6eb9-fb27-4731-bd91-b1b73a602287-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830411 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-bound-sa-token\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830458 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnmp5\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-kube-api-access-xnmp5\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.830483 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-tls\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.831299 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/80ff6eb9-fb27-4731-bd91-b1b73a602287-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.831636 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-certificates\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.832503 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ff6eb9-fb27-4731-bd91-b1b73a602287-trusted-ca\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.841262 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-registry-tls\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.841306 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/80ff6eb9-fb27-4731-bd91-b1b73a602287-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.846955 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-bound-sa-token\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.849024 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnmp5\" (UniqueName: \"kubernetes.io/projected/80ff6eb9-fb27-4731-bd91-b1b73a602287-kube-api-access-xnmp5\") pod \"image-registry-66df7c8f76-hl6gw\" (UID: \"80ff6eb9-fb27-4731-bd91-b1b73a602287\") " pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:04 crc kubenswrapper[4899]: I0123 08:57:04.890198 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:05 crc kubenswrapper[4899]: I0123 08:57:05.332560 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hl6gw"] Jan 23 08:57:05 crc kubenswrapper[4899]: W0123 08:57:05.338392 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80ff6eb9_fb27_4731_bd91_b1b73a602287.slice/crio-8a5af6994321c08b303f82885537729c70448a11334b962c457c8f04bbc97e31 WatchSource:0}: Error finding container 8a5af6994321c08b303f82885537729c70448a11334b962c457c8f04bbc97e31: Status 404 returned error can't find the container with id 8a5af6994321c08b303f82885537729c70448a11334b962c457c8f04bbc97e31 Jan 23 08:57:05 crc kubenswrapper[4899]: I0123 08:57:05.603660 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" event={"ID":"80ff6eb9-fb27-4731-bd91-b1b73a602287","Type":"ContainerStarted","Data":"dc112c68582119e058a9510a7d0b1fd41db470dce8d5d96b86223324e7a7493a"} Jan 23 08:57:05 crc kubenswrapper[4899]: I0123 08:57:05.603731 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" event={"ID":"80ff6eb9-fb27-4731-bd91-b1b73a602287","Type":"ContainerStarted","Data":"8a5af6994321c08b303f82885537729c70448a11334b962c457c8f04bbc97e31"} Jan 23 08:57:05 crc kubenswrapper[4899]: I0123 08:57:05.603873 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:05 crc kubenswrapper[4899]: I0123 08:57:05.621840 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" podStartSLOduration=1.621820779 podStartE2EDuration="1.621820779s" podCreationTimestamp="2026-01-23 08:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:57:05.619248713 +0000 UTC m=+312.486219120" watchObservedRunningTime="2026-01-23 08:57:05.621820779 +0000 UTC m=+312.488791166" Jan 23 08:57:24 crc kubenswrapper[4899]: I0123 08:57:24.895319 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hl6gw" Jan 23 08:57:24 crc kubenswrapper[4899]: I0123 08:57:24.948680 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.925660 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.926623 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zgmpt" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="registry-server" containerID="cri-o://62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc" gracePeriod=30 Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.936709 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.936991 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zcjjh" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="registry-server" containerID="cri-o://b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54" gracePeriod=30 Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.944816 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.945052 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" containerID="cri-o://86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662" gracePeriod=30 Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.959464 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.959742 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4ckf" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="registry-server" containerID="cri-o://f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1" gracePeriod=30 Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.979242 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.979515 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rstzc" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="registry-server" containerID="cri-o://15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d" gracePeriod=30 Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.986127 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-p45p8"] Jan 23 08:57:41 crc kubenswrapper[4899]: I0123 08:57:41.989710 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.019300 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-p45p8"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.061600 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4xc\" (UniqueName: \"kubernetes.io/projected/12cc03b0-6fca-4176-9a46-157b58c959d0-kube-api-access-9l4xc\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.061638 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.061811 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.163156 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4xc\" (UniqueName: \"kubernetes.io/projected/12cc03b0-6fca-4176-9a46-157b58c959d0-kube-api-access-9l4xc\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.163200 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.163255 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.165796 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.169838 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/12cc03b0-6fca-4176-9a46-157b58c959d0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.183840 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4xc\" (UniqueName: \"kubernetes.io/projected/12cc03b0-6fca-4176-9a46-157b58c959d0-kube-api-access-9l4xc\") pod \"marketplace-operator-79b997595-p45p8\" (UID: \"12cc03b0-6fca-4176-9a46-157b58c959d0\") " pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.340944 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.358426 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.405553 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.410699 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.412871 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.420496 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.467923 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content\") pod \"4625e21a-e288-4ab7-ac0b-b392f35778ac\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.467989 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74nhw\" (UniqueName: \"kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw\") pod \"4625e21a-e288-4ab7-ac0b-b392f35778ac\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.468116 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities\") pod \"4625e21a-e288-4ab7-ac0b-b392f35778ac\" (UID: \"4625e21a-e288-4ab7-ac0b-b392f35778ac\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.471648 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw" (OuterVolumeSpecName: "kube-api-access-74nhw") pod "4625e21a-e288-4ab7-ac0b-b392f35778ac" (UID: "4625e21a-e288-4ab7-ac0b-b392f35778ac"). InnerVolumeSpecName "kube-api-access-74nhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.472716 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities" (OuterVolumeSpecName: "utilities") pod "4625e21a-e288-4ab7-ac0b-b392f35778ac" (UID: "4625e21a-e288-4ab7-ac0b-b392f35778ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.522775 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4625e21a-e288-4ab7-ac0b-b392f35778ac" (UID: "4625e21a-e288-4ab7-ac0b-b392f35778ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569242 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content\") pod \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569307 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sw57\" (UniqueName: \"kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57\") pod \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569343 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities\") pod \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569377 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content\") pod \"a124473f-1d9e-40ff-99bb-5413ce650662\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569399 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prgrc\" (UniqueName: \"kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc\") pod \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569422 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities\") pod \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\" (UID: \"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569441 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcb4b\" (UniqueName: \"kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b\") pod \"a124473f-1d9e-40ff-99bb-5413ce650662\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569472 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content\") pod \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\" (UID: \"9c23cd89-e74f-499c-ab6b-fb76e68004b5\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569525 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities\") pod \"a124473f-1d9e-40ff-99bb-5413ce650662\" (UID: \"a124473f-1d9e-40ff-99bb-5413ce650662\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569554 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") pod \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569575 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") pod \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569594 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6p5h\" (UniqueName: \"kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h\") pod \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\" (UID: \"7d51b363-c51f-4ccd-b7ca-77ad850024dc\") " Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569791 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569810 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74nhw\" (UniqueName: \"kubernetes.io/projected/4625e21a-e288-4ab7-ac0b-b392f35778ac-kube-api-access-74nhw\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.569820 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4625e21a-e288-4ab7-ac0b-b392f35778ac-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.571075 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities" (OuterVolumeSpecName: "utilities") pod "9c23cd89-e74f-499c-ab6b-fb76e68004b5" (UID: "9c23cd89-e74f-499c-ab6b-fb76e68004b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.571635 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7d51b363-c51f-4ccd-b7ca-77ad850024dc" (UID: "7d51b363-c51f-4ccd-b7ca-77ad850024dc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.572221 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities" (OuterVolumeSpecName: "utilities") pod "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" (UID: "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.572996 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57" (OuterVolumeSpecName: "kube-api-access-9sw57") pod "9c23cd89-e74f-499c-ab6b-fb76e68004b5" (UID: "9c23cd89-e74f-499c-ab6b-fb76e68004b5"). InnerVolumeSpecName "kube-api-access-9sw57". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.574034 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc" (OuterVolumeSpecName: "kube-api-access-prgrc") pod "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" (UID: "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50"). InnerVolumeSpecName "kube-api-access-prgrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.574170 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7d51b363-c51f-4ccd-b7ca-77ad850024dc" (UID: "7d51b363-c51f-4ccd-b7ca-77ad850024dc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.578028 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b" (OuterVolumeSpecName: "kube-api-access-pcb4b") pod "a124473f-1d9e-40ff-99bb-5413ce650662" (UID: "a124473f-1d9e-40ff-99bb-5413ce650662"). InnerVolumeSpecName "kube-api-access-pcb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.581838 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities" (OuterVolumeSpecName: "utilities") pod "a124473f-1d9e-40ff-99bb-5413ce650662" (UID: "a124473f-1d9e-40ff-99bb-5413ce650662"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.582326 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h" (OuterVolumeSpecName: "kube-api-access-s6p5h") pod "7d51b363-c51f-4ccd-b7ca-77ad850024dc" (UID: "7d51b363-c51f-4ccd-b7ca-77ad850024dc"). InnerVolumeSpecName "kube-api-access-s6p5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.592786 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a124473f-1d9e-40ff-99bb-5413ce650662" (UID: "a124473f-1d9e-40ff-99bb-5413ce650662"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.629265 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" (UID: "e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671006 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sw57\" (UniqueName: \"kubernetes.io/projected/9c23cd89-e74f-499c-ab6b-fb76e68004b5-kube-api-access-9sw57\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671080 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671103 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671116 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prgrc\" (UniqueName: \"kubernetes.io/projected/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-kube-api-access-prgrc\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671128 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671167 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcb4b\" (UniqueName: \"kubernetes.io/projected/a124473f-1d9e-40ff-99bb-5413ce650662-kube-api-access-pcb4b\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671179 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a124473f-1d9e-40ff-99bb-5413ce650662-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671194 4899 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671244 4899 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7d51b363-c51f-4ccd-b7ca-77ad850024dc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671262 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6p5h\" (UniqueName: \"kubernetes.io/projected/7d51b363-c51f-4ccd-b7ca-77ad850024dc-kube-api-access-s6p5h\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.671274 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.694071 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c23cd89-e74f-499c-ab6b-fb76e68004b5" (UID: "9c23cd89-e74f-499c-ab6b-fb76e68004b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.748754 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-p45p8"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.772366 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c23cd89-e74f-499c-ab6b-fb76e68004b5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.812595 4899 generic.go:334] "Generic (PLEG): container finished" podID="a124473f-1d9e-40ff-99bb-5413ce650662" containerID="f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1" exitCode=0 Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.812654 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ckf" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.812665 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerDied","Data":"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.812697 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ckf" event={"ID":"a124473f-1d9e-40ff-99bb-5413ce650662","Type":"ContainerDied","Data":"92e511a5f80f33424bfbaa8e7e26219fb0464a98551606dd1cf2f42e92f7bbb4"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.812718 4899 scope.go:117] "RemoveContainer" containerID="f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.813520 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" event={"ID":"12cc03b0-6fca-4176-9a46-157b58c959d0","Type":"ContainerStarted","Data":"181f17e50a97ad16f104e7e6155803a53a748a8eb6e30f1d8041e6d79b6a3164"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.818238 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerDied","Data":"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.818185 4899 generic.go:334] "Generic (PLEG): container finished" podID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerID="86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662" exitCode=0 Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.818326 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" event={"ID":"7d51b363-c51f-4ccd-b7ca-77ad850024dc","Type":"ContainerDied","Data":"7b395f2ab6583378458029cff2a5a8592d70fa9f6c13f1a628221c26d0710e08"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.818421 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.824850 4899 generic.go:334] "Generic (PLEG): container finished" podID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerID="15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d" exitCode=0 Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.824937 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rstzc" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.824941 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerDied","Data":"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.824985 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rstzc" event={"ID":"9c23cd89-e74f-499c-ab6b-fb76e68004b5","Type":"ContainerDied","Data":"ade66961204c8cc0562aa5b2cc1258f4b98c3204b4f460536ed9316735f7978c"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.827419 4899 generic.go:334] "Generic (PLEG): container finished" podID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerID="b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54" exitCode=0 Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.827453 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerDied","Data":"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.827469 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcjjh" event={"ID":"e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50","Type":"ContainerDied","Data":"027879a8c5b62bb13d9ccdcc482529e19585bbbd26b715f8f85d8e6c7cce631e"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.827526 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcjjh" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.831616 4899 generic.go:334] "Generic (PLEG): container finished" podID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerID="62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc" exitCode=0 Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.831670 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerDied","Data":"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.831688 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zgmpt" event={"ID":"4625e21a-e288-4ab7-ac0b-b392f35778ac","Type":"ContainerDied","Data":"cc9034bd894bd51eb5b2440a3bca949000815fd9243f4a74edf2d322064bf982"} Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.831757 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zgmpt" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.835880 4899 scope.go:117] "RemoveContainer" containerID="0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.859414 4899 scope.go:117] "RemoveContainer" containerID="4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.870882 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.874258 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ckf"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.880673 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.888312 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zcjjh"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.891192 4899 scope.go:117] "RemoveContainer" containerID="f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.892108 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1\": container with ID starting with f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1 not found: ID does not exist" containerID="f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892148 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1"} err="failed to get container status \"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1\": rpc error: code = NotFound desc = could not find container \"f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1\": container with ID starting with f6a2583405e2f452fb49d6c5867f40002e1a7e9e36f401326b693c2e670171b1 not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892175 4899 scope.go:117] "RemoveContainer" containerID="0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.892483 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a\": container with ID starting with 0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a not found: ID does not exist" containerID="0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892523 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a"} err="failed to get container status \"0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a\": rpc error: code = NotFound desc = could not find container \"0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a\": container with ID starting with 0722934c1b8523fda2470c21b2cd87f0e6485b6189dfb2154d500f3cb76ab90a not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892541 4899 scope.go:117] "RemoveContainer" containerID="4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892651 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.892810 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9\": container with ID starting with 4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9 not found: ID does not exist" containerID="4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892834 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9"} err="failed to get container status \"4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9\": rpc error: code = NotFound desc = could not find container \"4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9\": container with ID starting with 4ba3f31758c0a888ab8af5d7218b0dbb67684bbb19eaa054e3de0189ca5c86c9 not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.892850 4899 scope.go:117] "RemoveContainer" containerID="86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.907261 4899 scope.go:117] "RemoveContainer" containerID="d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.908832 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rstzc"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.914539 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.919111 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zgmpt"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.923504 4899 scope.go:117] "RemoveContainer" containerID="86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.923786 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.923968 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662\": container with ID starting with 86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662 not found: ID does not exist" containerID="86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.923998 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662"} err="failed to get container status \"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662\": rpc error: code = NotFound desc = could not find container \"86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662\": container with ID starting with 86845360e5ce002aea616f642b02dbfbe0164556171dc48357faa2669604b662 not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.924030 4899 scope.go:117] "RemoveContainer" containerID="d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.924552 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd\": container with ID starting with d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd not found: ID does not exist" containerID="d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.924595 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd"} err="failed to get container status \"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd\": rpc error: code = NotFound desc = could not find container \"d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd\": container with ID starting with d88be743476d20b4835949885c5707585ab9664be513f0fc8cf9cb40994461bd not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.924614 4899 scope.go:117] "RemoveContainer" containerID="15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.928666 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wcqlx"] Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.936511 4899 scope.go:117] "RemoveContainer" containerID="cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.950946 4899 scope.go:117] "RemoveContainer" containerID="f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.966245 4899 scope.go:117] "RemoveContainer" containerID="15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.966711 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d\": container with ID starting with 15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d not found: ID does not exist" containerID="15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.966762 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d"} err="failed to get container status \"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d\": rpc error: code = NotFound desc = could not find container \"15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d\": container with ID starting with 15e1a243253e8a9d6210103252e2c0a5d71c0e70125750d92e642c6851edea8d not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.966792 4899 scope.go:117] "RemoveContainer" containerID="cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.967299 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7\": container with ID starting with cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7 not found: ID does not exist" containerID="cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.967339 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7"} err="failed to get container status \"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7\": rpc error: code = NotFound desc = could not find container \"cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7\": container with ID starting with cffa3fa9faa48ad6631c220e8f548e5be55fd20fa015c965a833ff6b882b95c7 not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.967367 4899 scope.go:117] "RemoveContainer" containerID="f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534" Jan 23 08:57:42 crc kubenswrapper[4899]: E0123 08:57:42.967789 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534\": container with ID starting with f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534 not found: ID does not exist" containerID="f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.967813 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534"} err="failed to get container status \"f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534\": rpc error: code = NotFound desc = could not find container \"f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534\": container with ID starting with f0f87e57194a8a40bbea46f52cd8653eafec11f05c1474b89f1c94f407769534 not found: ID does not exist" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.967832 4899 scope.go:117] "RemoveContainer" containerID="b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.984480 4899 scope.go:117] "RemoveContainer" containerID="7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d" Jan 23 08:57:42 crc kubenswrapper[4899]: I0123 08:57:42.998890 4899 scope.go:117] "RemoveContainer" containerID="84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.026517 4899 scope.go:117] "RemoveContainer" containerID="b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.027134 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54\": container with ID starting with b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54 not found: ID does not exist" containerID="b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027174 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54"} err="failed to get container status \"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54\": rpc error: code = NotFound desc = could not find container \"b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54\": container with ID starting with b8e4aff0fe6dbf774209a24a1948b1bf416c03bd52f636899b36a7ca1dff7d54 not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027224 4899 scope.go:117] "RemoveContainer" containerID="7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.027534 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d\": container with ID starting with 7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d not found: ID does not exist" containerID="7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027555 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d"} err="failed to get container status \"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d\": rpc error: code = NotFound desc = could not find container \"7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d\": container with ID starting with 7027672b24223619681a57ba0092cb26e059d4f91ee421358e3c23a89ae7235d not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027568 4899 scope.go:117] "RemoveContainer" containerID="84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.027901 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70\": container with ID starting with 84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70 not found: ID does not exist" containerID="84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027922 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70"} err="failed to get container status \"84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70\": rpc error: code = NotFound desc = could not find container \"84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70\": container with ID starting with 84becd246235fde60c4825cd788a79bf1a1c18d5744baf5243ee3dd147478a70 not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.027937 4899 scope.go:117] "RemoveContainer" containerID="62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.045566 4899 scope.go:117] "RemoveContainer" containerID="4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.062066 4899 scope.go:117] "RemoveContainer" containerID="892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.077507 4899 scope.go:117] "RemoveContainer" containerID="62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.077903 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc\": container with ID starting with 62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc not found: ID does not exist" containerID="62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.077931 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc"} err="failed to get container status \"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc\": rpc error: code = NotFound desc = could not find container \"62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc\": container with ID starting with 62340f776e55b12900dbd6f9427c0ef8fe399ff5caa50e17a2bc126b7be7c9dc not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.077953 4899 scope.go:117] "RemoveContainer" containerID="4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.079008 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc\": container with ID starting with 4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc not found: ID does not exist" containerID="4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.079041 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc"} err="failed to get container status \"4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc\": rpc error: code = NotFound desc = could not find container \"4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc\": container with ID starting with 4059d5f6baa4fbcce4218bec1d3ed03f25a6bce3ae4cb24cb954633edd37cfdc not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.079057 4899 scope.go:117] "RemoveContainer" containerID="892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e" Jan 23 08:57:43 crc kubenswrapper[4899]: E0123 08:57:43.079432 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e\": container with ID starting with 892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e not found: ID does not exist" containerID="892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.079452 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e"} err="failed to get container status \"892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e\": rpc error: code = NotFound desc = could not find container \"892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e\": container with ID starting with 892c6be6a2670517eca78d986152002211a8afb415f7725901498c00c87e761e not found: ID does not exist" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.226371 4899 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wcqlx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.226468 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wcqlx" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.524556 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" path="/var/lib/kubelet/pods/4625e21a-e288-4ab7-ac0b-b392f35778ac/volumes" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.525936 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" path="/var/lib/kubelet/pods/7d51b363-c51f-4ccd-b7ca-77ad850024dc/volumes" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.528884 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" path="/var/lib/kubelet/pods/9c23cd89-e74f-499c-ab6b-fb76e68004b5/volumes" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.529567 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" path="/var/lib/kubelet/pods/a124473f-1d9e-40ff-99bb-5413ce650662/volumes" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.530696 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" path="/var/lib/kubelet/pods/e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50/volumes" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.839055 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" event={"ID":"12cc03b0-6fca-4176-9a46-157b58c959d0","Type":"ContainerStarted","Data":"b4fc56767d13b62eec0f070d80af1f61f963acd730f7ea3eab2c7b6bf5206816"} Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.839483 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.842259 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" Jan 23 08:57:43 crc kubenswrapper[4899]: I0123 08:57:43.878876 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-p45p8" podStartSLOduration=2.8788606469999998 podStartE2EDuration="2.878860647s" podCreationTimestamp="2026-01-23 08:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 08:57:43.859142854 +0000 UTC m=+350.726113241" watchObservedRunningTime="2026-01-23 08:57:43.878860647 +0000 UTC m=+350.745831034" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.358605 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359351 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359365 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359377 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359383 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359391 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359397 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359405 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359411 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359419 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359425 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359433 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359439 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359448 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359453 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359460 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359465 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359474 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359479 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359486 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359492 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="extract-content" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359507 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359514 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359523 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359529 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="extract-utilities" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359538 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359545 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: E0123 08:57:46.359556 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359563 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359658 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74c9e8d-7ce9-44fc-b33d-b6fc33d31a50" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359670 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="a124473f-1d9e-40ff-99bb-5413ce650662" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359681 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="4625e21a-e288-4ab7-ac0b-b392f35778ac" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359699 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359711 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d51b363-c51f-4ccd-b7ca-77ad850024dc" containerName="marketplace-operator" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.359726 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c23cd89-e74f-499c-ab6b-fb76e68004b5" containerName="registry-server" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.360570 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.362886 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.371920 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.525020 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.525383 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.525472 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn2fs\" (UniqueName: \"kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.558681 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q5nf4"] Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.559894 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.564049 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.577865 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q5nf4"] Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.627161 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.627256 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.627296 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn2fs\" (UniqueName: \"kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.627798 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.627986 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.647638 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn2fs\" (UniqueName: \"kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs\") pod \"certified-operators-vtljc\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.681596 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.728167 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-utilities\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.728224 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-catalog-content\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.728296 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4s5s\" (UniqueName: \"kubernetes.io/projected/683effa2-db47-4531-9aad-da9b77a8daf9-kube-api-access-w4s5s\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.830200 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4s5s\" (UniqueName: \"kubernetes.io/projected/683effa2-db47-4531-9aad-da9b77a8daf9-kube-api-access-w4s5s\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.830289 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-utilities\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.830314 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-catalog-content\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.831412 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-utilities\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.832235 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683effa2-db47-4531-9aad-da9b77a8daf9-catalog-content\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.850880 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4s5s\" (UniqueName: \"kubernetes.io/projected/683effa2-db47-4531-9aad-da9b77a8daf9-kube-api-access-w4s5s\") pod \"community-operators-q5nf4\" (UID: \"683effa2-db47-4531-9aad-da9b77a8daf9\") " pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:46 crc kubenswrapper[4899]: I0123 08:57:46.889811 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.120037 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 08:57:47 crc kubenswrapper[4899]: W0123 08:57:47.123609 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda03a9c2b_6cd4_4ff6_9085_121a461bf730.slice/crio-34646bcc3c4d68f1fc010658d60f6cc7494c2226fe1a37d19096623807f58d55 WatchSource:0}: Error finding container 34646bcc3c4d68f1fc010658d60f6cc7494c2226fe1a37d19096623807f58d55: Status 404 returned error can't find the container with id 34646bcc3c4d68f1fc010658d60f6cc7494c2226fe1a37d19096623807f58d55 Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.279751 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q5nf4"] Jan 23 08:57:47 crc kubenswrapper[4899]: W0123 08:57:47.334107 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod683effa2_db47_4531_9aad_da9b77a8daf9.slice/crio-8643edf7ec85001a89947cb07bbee95398863c4a12dd758c996f43c7b527bd58 WatchSource:0}: Error finding container 8643edf7ec85001a89947cb07bbee95398863c4a12dd758c996f43c7b527bd58: Status 404 returned error can't find the container with id 8643edf7ec85001a89947cb07bbee95398863c4a12dd758c996f43c7b527bd58 Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.865594 4899 generic.go:334] "Generic (PLEG): container finished" podID="683effa2-db47-4531-9aad-da9b77a8daf9" containerID="9a3375cf5b823a3ac60d323d2cc00460db1128f67b6490ea71f7764e40c75f2e" exitCode=0 Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.866023 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5nf4" event={"ID":"683effa2-db47-4531-9aad-da9b77a8daf9","Type":"ContainerDied","Data":"9a3375cf5b823a3ac60d323d2cc00460db1128f67b6490ea71f7764e40c75f2e"} Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.866068 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5nf4" event={"ID":"683effa2-db47-4531-9aad-da9b77a8daf9","Type":"ContainerStarted","Data":"8643edf7ec85001a89947cb07bbee95398863c4a12dd758c996f43c7b527bd58"} Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.868738 4899 generic.go:334] "Generic (PLEG): container finished" podID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerID="acd89e1851a2d5566894f2879d3905d4bcf847232224ba544fa54cc4d44a0f7a" exitCode=0 Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.868771 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerDied","Data":"acd89e1851a2d5566894f2879d3905d4bcf847232224ba544fa54cc4d44a0f7a"} Jan 23 08:57:47 crc kubenswrapper[4899]: I0123 08:57:47.868797 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerStarted","Data":"34646bcc3c4d68f1fc010658d60f6cc7494c2226fe1a37d19096623807f58d55"} Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.763015 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.765759 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.768644 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.772695 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.862255 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrzbn\" (UniqueName: \"kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.862334 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.862441 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.877974 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5nf4" event={"ID":"683effa2-db47-4531-9aad-da9b77a8daf9","Type":"ContainerStarted","Data":"4f06902b1274d415ea480fc943cf2b2984c3929d703dba84e057fb79d9d703a9"} Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.879931 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerStarted","Data":"896bbdf82fe34a8e0916cde513a6cef6bc2e9a375f2e8f902831ec6c50ad7519"} Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.963151 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.963257 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrzbn\" (UniqueName: \"kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.963285 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.963472 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fq4bf"] Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.964163 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.964168 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.964678 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.967578 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.969494 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fq4bf"] Jan 23 08:57:48 crc kubenswrapper[4899]: I0123 08:57:48.986221 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrzbn\" (UniqueName: \"kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn\") pod \"redhat-marketplace-4x76s\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.064298 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-utilities\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.064371 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsmdq\" (UniqueName: \"kubernetes.io/projected/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-kube-api-access-jsmdq\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.064398 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-catalog-content\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.082909 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.165028 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsmdq\" (UniqueName: \"kubernetes.io/projected/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-kube-api-access-jsmdq\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.165077 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-catalog-content\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.165126 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-utilities\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.165578 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-catalog-content\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.165610 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-utilities\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.187842 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsmdq\" (UniqueName: \"kubernetes.io/projected/cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56-kube-api-access-jsmdq\") pod \"redhat-operators-fq4bf\" (UID: \"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56\") " pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.281867 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.451972 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fq4bf"] Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.480783 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 08:57:49 crc kubenswrapper[4899]: W0123 08:57:49.481876 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d620e98_4c84_4ff3_a311_c839033909aa.slice/crio-56a5c3ddccbde34946c8a8b7bb764cde9b4d16b397839716a356df486df5127d WatchSource:0}: Error finding container 56a5c3ddccbde34946c8a8b7bb764cde9b4d16b397839716a356df486df5127d: Status 404 returned error can't find the container with id 56a5c3ddccbde34946c8a8b7bb764cde9b4d16b397839716a356df486df5127d Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.888838 4899 generic.go:334] "Generic (PLEG): container finished" podID="1d620e98-4c84-4ff3-a311-c839033909aa" containerID="b804b2a710ef73a83d124ade32ff344d36a739d0bcf4182c3f3df2d1c3915cae" exitCode=0 Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.888969 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerDied","Data":"b804b2a710ef73a83d124ade32ff344d36a739d0bcf4182c3f3df2d1c3915cae"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.889030 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerStarted","Data":"56a5c3ddccbde34946c8a8b7bb764cde9b4d16b397839716a356df486df5127d"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.892680 4899 generic.go:334] "Generic (PLEG): container finished" podID="683effa2-db47-4531-9aad-da9b77a8daf9" containerID="4f06902b1274d415ea480fc943cf2b2984c3929d703dba84e057fb79d9d703a9" exitCode=0 Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.892715 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5nf4" event={"ID":"683effa2-db47-4531-9aad-da9b77a8daf9","Type":"ContainerDied","Data":"4f06902b1274d415ea480fc943cf2b2984c3929d703dba84e057fb79d9d703a9"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.898462 4899 generic.go:334] "Generic (PLEG): container finished" podID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerID="896bbdf82fe34a8e0916cde513a6cef6bc2e9a375f2e8f902831ec6c50ad7519" exitCode=0 Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.898585 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerDied","Data":"896bbdf82fe34a8e0916cde513a6cef6bc2e9a375f2e8f902831ec6c50ad7519"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.900320 4899 generic.go:334] "Generic (PLEG): container finished" podID="cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56" containerID="45151db8f6c5e88ac0e7ee464893e9b37febe8ad329c4ac12ddafe3440495e88" exitCode=0 Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.900368 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq4bf" event={"ID":"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56","Type":"ContainerDied","Data":"45151db8f6c5e88ac0e7ee464893e9b37febe8ad329c4ac12ddafe3440495e88"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.900388 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq4bf" event={"ID":"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56","Type":"ContainerStarted","Data":"1ee957de2f8cb79140e76185813da07b6640878edf4ad6be8f7fe1cffaefd2a0"} Jan 23 08:57:49 crc kubenswrapper[4899]: I0123 08:57:49.992449 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" podUID="bc1b652a-534c-414e-86b3-00b89a82c21e" containerName="registry" containerID="cri-o://fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4" gracePeriod=30 Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.384620 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.482786 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.482929 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.482964 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.482995 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483015 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483060 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483079 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483138 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjmp7\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7\") pod \"bc1b652a-534c-414e-86b3-00b89a82c21e\" (UID: \"bc1b652a-534c-414e-86b3-00b89a82c21e\") " Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483970 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.483991 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.489700 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.492993 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.495548 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7" (OuterVolumeSpecName: "kube-api-access-sjmp7") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "kube-api-access-sjmp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.495678 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.495768 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.504526 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bc1b652a-534c-414e-86b3-00b89a82c21e" (UID: "bc1b652a-534c-414e-86b3-00b89a82c21e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585831 4899 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585870 4899 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bc1b652a-534c-414e-86b3-00b89a82c21e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585880 4899 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc1b652a-534c-414e-86b3-00b89a82c21e-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585889 4899 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585898 4899 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bc1b652a-534c-414e-86b3-00b89a82c21e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585906 4899 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.585913 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjmp7\" (UniqueName: \"kubernetes.io/projected/bc1b652a-534c-414e-86b3-00b89a82c21e-kube-api-access-sjmp7\") on node \"crc\" DevicePath \"\"" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.906725 4899 generic.go:334] "Generic (PLEG): container finished" podID="1d620e98-4c84-4ff3-a311-c839033909aa" containerID="f98bac878649c293ca9f762a7f3256ec30005e4f09d0270f648749d8e192b7ec" exitCode=0 Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.907020 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerDied","Data":"f98bac878649c293ca9f762a7f3256ec30005e4f09d0270f648749d8e192b7ec"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.909719 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5nf4" event={"ID":"683effa2-db47-4531-9aad-da9b77a8daf9","Type":"ContainerStarted","Data":"700dcae2e3d490949676e5e36c6693f53e3d9e64d5ef16771422e999eb74670e"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.912358 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerStarted","Data":"ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.913852 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq4bf" event={"ID":"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56","Type":"ContainerStarted","Data":"c4d685633e2cc4495e959346cf0275999e6e62a0f821ac2f2536228342c4f582"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.915423 4899 generic.go:334] "Generic (PLEG): container finished" podID="bc1b652a-534c-414e-86b3-00b89a82c21e" containerID="fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4" exitCode=0 Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.915452 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" event={"ID":"bc1b652a-534c-414e-86b3-00b89a82c21e","Type":"ContainerDied","Data":"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.915467 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" event={"ID":"bc1b652a-534c-414e-86b3-00b89a82c21e","Type":"ContainerDied","Data":"b616b35e36af780631ba6464a79f5cbbe16c4c3a725134ade56d987901b6afca"} Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.915484 4899 scope.go:117] "RemoveContainer" containerID="fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.915585 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l2tzs" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.936838 4899 scope.go:117] "RemoveContainer" containerID="fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4" Jan 23 08:57:50 crc kubenswrapper[4899]: E0123 08:57:50.937333 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4\": container with ID starting with fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4 not found: ID does not exist" containerID="fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.937393 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4"} err="failed to get container status \"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4\": rpc error: code = NotFound desc = could not find container \"fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4\": container with ID starting with fc7427b82c1430c8f3858ba6c86f15d82fcf1df162bc34bbe5f3fdf88440b7c4 not found: ID does not exist" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.975702 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q5nf4" podStartSLOduration=2.486022484 podStartE2EDuration="4.97568594s" podCreationTimestamp="2026-01-23 08:57:46 +0000 UTC" firstStartedPulling="2026-01-23 08:57:47.868879469 +0000 UTC m=+354.735849856" lastFinishedPulling="2026-01-23 08:57:50.358542925 +0000 UTC m=+357.225513312" observedRunningTime="2026-01-23 08:57:50.955171974 +0000 UTC m=+357.822142371" watchObservedRunningTime="2026-01-23 08:57:50.97568594 +0000 UTC m=+357.842656327" Jan 23 08:57:50 crc kubenswrapper[4899]: I0123 08:57:50.976202 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vtljc" podStartSLOduration=2.420225808 podStartE2EDuration="4.976197194s" podCreationTimestamp="2026-01-23 08:57:46 +0000 UTC" firstStartedPulling="2026-01-23 08:57:47.869588449 +0000 UTC m=+354.736558836" lastFinishedPulling="2026-01-23 08:57:50.425559835 +0000 UTC m=+357.292530222" observedRunningTime="2026-01-23 08:57:50.973880089 +0000 UTC m=+357.840850496" watchObservedRunningTime="2026-01-23 08:57:50.976197194 +0000 UTC m=+357.843167591" Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.012124 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.019630 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l2tzs"] Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.523715 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc1b652a-534c-414e-86b3-00b89a82c21e" path="/var/lib/kubelet/pods/bc1b652a-534c-414e-86b3-00b89a82c21e/volumes" Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.924639 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerStarted","Data":"6fb0fa33f1e78ae4eea1410da40a74a18ab6cb1e9b1831dabc3d616ee581091e"} Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.927009 4899 generic.go:334] "Generic (PLEG): container finished" podID="cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56" containerID="c4d685633e2cc4495e959346cf0275999e6e62a0f821ac2f2536228342c4f582" exitCode=0 Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.927139 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq4bf" event={"ID":"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56","Type":"ContainerDied","Data":"c4d685633e2cc4495e959346cf0275999e6e62a0f821ac2f2536228342c4f582"} Jan 23 08:57:51 crc kubenswrapper[4899]: I0123 08:57:51.951302 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4x76s" podStartSLOduration=2.438559159 podStartE2EDuration="3.951286443s" podCreationTimestamp="2026-01-23 08:57:48 +0000 UTC" firstStartedPulling="2026-01-23 08:57:49.893785724 +0000 UTC m=+356.760756111" lastFinishedPulling="2026-01-23 08:57:51.406513008 +0000 UTC m=+358.273483395" observedRunningTime="2026-01-23 08:57:51.948426773 +0000 UTC m=+358.815397160" watchObservedRunningTime="2026-01-23 08:57:51.951286443 +0000 UTC m=+358.818256830" Jan 23 08:57:52 crc kubenswrapper[4899]: I0123 08:57:52.935723 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq4bf" event={"ID":"cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56","Type":"ContainerStarted","Data":"72c4765b13666c758ee67fb6ff6a24ce9a8c68a6da37a9feebd944e0dc231677"} Jan 23 08:57:52 crc kubenswrapper[4899]: I0123 08:57:52.966865 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fq4bf" podStartSLOduration=2.560029268 podStartE2EDuration="4.966845768s" podCreationTimestamp="2026-01-23 08:57:48 +0000 UTC" firstStartedPulling="2026-01-23 08:57:49.901775139 +0000 UTC m=+356.768745526" lastFinishedPulling="2026-01-23 08:57:52.308591639 +0000 UTC m=+359.175562026" observedRunningTime="2026-01-23 08:57:52.966418366 +0000 UTC m=+359.833388753" watchObservedRunningTime="2026-01-23 08:57:52.966845768 +0000 UTC m=+359.833816155" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.681723 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.682278 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.723377 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.890754 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.890802 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.927722 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:56 crc kubenswrapper[4899]: I0123 08:57:56.984834 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 08:57:57 crc kubenswrapper[4899]: I0123 08:57:57.056671 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q5nf4" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.083224 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.083569 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.122081 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.283252 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.283303 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:57:59 crc kubenswrapper[4899]: I0123 08:57:59.327110 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:58:00 crc kubenswrapper[4899]: I0123 08:58:00.001986 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fq4bf" Jan 23 08:58:00 crc kubenswrapper[4899]: I0123 08:58:00.002709 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 08:58:01 crc kubenswrapper[4899]: I0123 08:58:01.685401 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:58:01 crc kubenswrapper[4899]: I0123 08:58:01.685749 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:58:31 crc kubenswrapper[4899]: I0123 08:58:31.685947 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:58:31 crc kubenswrapper[4899]: I0123 08:58:31.686714 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:59:01 crc kubenswrapper[4899]: I0123 08:59:01.686264 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 08:59:01 crc kubenswrapper[4899]: I0123 08:59:01.686922 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 08:59:01 crc kubenswrapper[4899]: I0123 08:59:01.687005 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 08:59:01 crc kubenswrapper[4899]: I0123 08:59:01.688044 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 08:59:01 crc kubenswrapper[4899]: I0123 08:59:01.688132 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b" gracePeriod=600 Jan 23 08:59:02 crc kubenswrapper[4899]: I0123 08:59:02.324872 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b" exitCode=0 Jan 23 08:59:02 crc kubenswrapper[4899]: I0123 08:59:02.324938 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b"} Jan 23 08:59:02 crc kubenswrapper[4899]: I0123 08:59:02.325371 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9"} Jan 23 08:59:02 crc kubenswrapper[4899]: I0123 08:59:02.325412 4899 scope.go:117] "RemoveContainer" containerID="e223db551b5e7fb87b7a0a795c413c2f672ba9dccb6b1ad03506434348448d5a" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.194863 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4"] Jan 23 09:00:00 crc kubenswrapper[4899]: E0123 09:00:00.195659 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1b652a-534c-414e-86b3-00b89a82c21e" containerName="registry" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.195674 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1b652a-534c-414e-86b3-00b89a82c21e" containerName="registry" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.195793 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1b652a-534c-414e-86b3-00b89a82c21e" containerName="registry" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.196307 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.200038 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.200133 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.203093 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4"] Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.299139 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2dv\" (UniqueName: \"kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.299253 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.299293 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.399887 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.399966 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.400017 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2dv\" (UniqueName: \"kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.401131 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.405903 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.418723 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2dv\" (UniqueName: \"kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv\") pod \"collect-profiles-29485980-tnpm4\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.514528 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:00 crc kubenswrapper[4899]: I0123 09:00:00.685080 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4"] Jan 23 09:00:01 crc kubenswrapper[4899]: I0123 09:00:01.649117 4899 generic.go:334] "Generic (PLEG): container finished" podID="4b345c17-887a-4970-9cf3-e4e3b59bbb94" containerID="e7c3053af8164f7b40f6931d42c9b9abe6695b60cb8acecedfc9a2f11b727c19" exitCode=0 Jan 23 09:00:01 crc kubenswrapper[4899]: I0123 09:00:01.649156 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" event={"ID":"4b345c17-887a-4970-9cf3-e4e3b59bbb94","Type":"ContainerDied","Data":"e7c3053af8164f7b40f6931d42c9b9abe6695b60cb8acecedfc9a2f11b727c19"} Jan 23 09:00:01 crc kubenswrapper[4899]: I0123 09:00:01.649188 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" event={"ID":"4b345c17-887a-4970-9cf3-e4e3b59bbb94","Type":"ContainerStarted","Data":"5544ec90441e156432526e6b7025defa833f1177038cb90abbaf9bfdd4b92dc7"} Jan 23 09:00:02 crc kubenswrapper[4899]: I0123 09:00:02.842785 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.035581 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx2dv\" (UniqueName: \"kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv\") pod \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.035661 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume\") pod \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.035721 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume\") pod \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\" (UID: \"4b345c17-887a-4970-9cf3-e4e3b59bbb94\") " Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.036455 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b345c17-887a-4970-9cf3-e4e3b59bbb94" (UID: "4b345c17-887a-4970-9cf3-e4e3b59bbb94"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.040640 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b345c17-887a-4970-9cf3-e4e3b59bbb94" (UID: "4b345c17-887a-4970-9cf3-e4e3b59bbb94"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.041381 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv" (OuterVolumeSpecName: "kube-api-access-zx2dv") pod "4b345c17-887a-4970-9cf3-e4e3b59bbb94" (UID: "4b345c17-887a-4970-9cf3-e4e3b59bbb94"). InnerVolumeSpecName "kube-api-access-zx2dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.136779 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx2dv\" (UniqueName: \"kubernetes.io/projected/4b345c17-887a-4970-9cf3-e4e3b59bbb94-kube-api-access-zx2dv\") on node \"crc\" DevicePath \"\"" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.136814 4899 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b345c17-887a-4970-9cf3-e4e3b59bbb94-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.136826 4899 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b345c17-887a-4970-9cf3-e4e3b59bbb94-config-volume\") on node \"crc\" DevicePath \"\"" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.661350 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" event={"ID":"4b345c17-887a-4970-9cf3-e4e3b59bbb94","Type":"ContainerDied","Data":"5544ec90441e156432526e6b7025defa833f1177038cb90abbaf9bfdd4b92dc7"} Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.661706 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5544ec90441e156432526e6b7025defa833f1177038cb90abbaf9bfdd4b92dc7" Jan 23 09:00:03 crc kubenswrapper[4899]: I0123 09:00:03.661419 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485980-tnpm4" Jan 23 09:01:01 crc kubenswrapper[4899]: I0123 09:01:01.686356 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:01:01 crc kubenswrapper[4899]: I0123 09:01:01.686855 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:01:31 crc kubenswrapper[4899]: I0123 09:01:31.685869 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:01:31 crc kubenswrapper[4899]: I0123 09:01:31.686627 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:02:01 crc kubenswrapper[4899]: I0123 09:02:01.686550 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:02:01 crc kubenswrapper[4899]: I0123 09:02:01.687306 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:02:01 crc kubenswrapper[4899]: I0123 09:02:01.687382 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:02:01 crc kubenswrapper[4899]: I0123 09:02:01.688519 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:02:01 crc kubenswrapper[4899]: I0123 09:02:01.688642 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9" gracePeriod=600 Jan 23 09:02:02 crc kubenswrapper[4899]: I0123 09:02:02.352986 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9" exitCode=0 Jan 23 09:02:02 crc kubenswrapper[4899]: I0123 09:02:02.353053 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9"} Jan 23 09:02:02 crc kubenswrapper[4899]: I0123 09:02:02.353345 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4"} Jan 23 09:02:02 crc kubenswrapper[4899]: I0123 09:02:02.353387 4899 scope.go:117] "RemoveContainer" containerID="61808d53602224cedb0527480bc58d715f88407c4050cdc72cb011382a18239b" Jan 23 09:04:01 crc kubenswrapper[4899]: I0123 09:04:01.685943 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:04:01 crc kubenswrapper[4899]: I0123 09:04:01.686543 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.161782 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qmqs"] Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163395 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-controller" containerID="cri-o://645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163460 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="nbdb" containerID="cri-o://ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163532 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="northd" containerID="cri-o://a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163574 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163604 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-node" containerID="cri-o://f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163637 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-acl-logging" containerID="cri-o://c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.163849 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="sbdb" containerID="cri-o://0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.189841 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" containerID="cri-o://ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" gracePeriod=30 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.295578 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/2.log" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.296201 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/1.log" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.296264 4899 generic.go:334] "Generic (PLEG): container finished" podID="60a84ebf-a188-4eee-8a9f-e02ac7adb8c8" containerID="f9014f9495c81d445c7b91ebf5490a70e7db4b2d956003d1991b6e84ed06afce" exitCode=2 Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.296312 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerDied","Data":"f9014f9495c81d445c7b91ebf5490a70e7db4b2d956003d1991b6e84ed06afce"} Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.296351 4899 scope.go:117] "RemoveContainer" containerID="cda169679407259a2ff88caba2eeb1dad9abdd32402fea2f9859dad24a8ed697" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.296877 4899 scope.go:117] "RemoveContainer" containerID="f9014f9495c81d445c7b91ebf5490a70e7db4b2d956003d1991b6e84ed06afce" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.688589 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.688641 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.690747 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.690857 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.691840 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.691879 4899 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="sbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.692862 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.692913 4899 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="nbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.932780 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/3.log" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.935920 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovn-acl-logging/0.log" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.936442 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovn-controller/0.log" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.936973 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998273 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kzbqx"] Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998516 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="northd" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998538 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="northd" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998555 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-acl-logging" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998566 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-acl-logging" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998577 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-ovn-metrics" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998585 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-ovn-metrics" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998599 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998605 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998614 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="nbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998620 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="nbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998628 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998636 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998645 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b345c17-887a-4970-9cf3-e4e3b59bbb94" containerName="collect-profiles" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998653 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b345c17-887a-4970-9cf3-e4e3b59bbb94" containerName="collect-profiles" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998666 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-node" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998673 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-node" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998684 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998691 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998702 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="sbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998709 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="sbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998722 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kubecfg-setup" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998729 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kubecfg-setup" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998737 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998744 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998755 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998763 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: E0123 09:04:18.998774 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998781 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998906 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="sbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998918 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-node" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998927 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998938 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="kube-rbac-proxy-ovn-metrics" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998947 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998954 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998963 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="nbdb" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998974 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998982 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b345c17-887a-4970-9cf3-e4e3b59bbb94" containerName="collect-profiles" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.998994 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="northd" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.999002 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-acl-logging" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.999011 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovn-controller" Jan 23 09:04:18 crc kubenswrapper[4899]: I0123 09:04:18.999225 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerName="ovnkube-controller" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.001007 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064081 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064359 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064430 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064499 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064562 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064628 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064691 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064761 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064830 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064757 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064777 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064786 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064813 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064819 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064833 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064856 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065077 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log" (OuterVolumeSpecName: "node-log") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065327 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065340 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.064898 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065829 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065905 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.065975 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066035 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066099 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066172 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash" (OuterVolumeSpecName: "host-slash") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066229 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066267 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066180 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066361 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066436 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066669 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066703 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mv9\" (UniqueName: \"kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066738 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066757 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch\") pod \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\" (UID: \"790411d3-0df3-4cc3-bf33-835dab4c2cb7\") " Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066807 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket" (OuterVolumeSpecName: "log-socket") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.066912 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067269 4899 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-slash\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067292 4899 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067302 4899 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067311 4899 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067319 4899 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067327 4899 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-log-socket\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067337 4899 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067346 4899 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067354 4899 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067362 4899 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067370 4899 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067378 4899 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067386 4899 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067395 4899 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067403 4899 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067411 4899 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/790411d3-0df3-4cc3-bf33-835dab4c2cb7-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.067420 4899 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-node-log\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.072139 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.073265 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9" (OuterVolumeSpecName: "kube-api-access-v4mv9") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "kube-api-access-v4mv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.080162 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "790411d3-0df3-4cc3-bf33-835dab4c2cb7" (UID: "790411d3-0df3-4cc3-bf33-835dab4c2cb7"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168144 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-ovn\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168480 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-netd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168513 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23888411-56af-4cb6-871f-4804b5bcebf7-ovn-node-metrics-cert\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168534 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168557 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-systemd-units\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168581 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-etc-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168603 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-env-overrides\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168669 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-kubelet\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168711 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168739 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-netns\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168771 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-log-socket\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168792 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-bin\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168813 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-systemd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168854 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-slash\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168882 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-node-log\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168898 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-var-lib-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168916 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-config\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168936 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-script-lib\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.168975 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4876n\" (UniqueName: \"kubernetes.io/projected/23888411-56af-4cb6-871f-4804b5bcebf7-kube-api-access-4876n\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.169003 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.169058 4899 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/790411d3-0df3-4cc3-bf33-835dab4c2cb7-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.169074 4899 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/790411d3-0df3-4cc3-bf33-835dab4c2cb7-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.169087 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mv9\" (UniqueName: \"kubernetes.io/projected/790411d3-0df3-4cc3-bf33-835dab4c2cb7-kube-api-access-v4mv9\") on node \"crc\" DevicePath \"\"" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269816 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4876n\" (UniqueName: \"kubernetes.io/projected/23888411-56af-4cb6-871f-4804b5bcebf7-kube-api-access-4876n\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269863 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269891 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-ovn\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269920 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-netd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269949 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23888411-56af-4cb6-871f-4804b5bcebf7-ovn-node-metrics-cert\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269970 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.269996 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-systemd-units\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270027 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-etc-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270046 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-env-overrides\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270074 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-kubelet\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270099 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270126 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-netns\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270177 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-log-socket\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270186 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270198 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-bin\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270249 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270261 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-systemd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270280 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-ovn\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270304 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-slash\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270324 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-netd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270333 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-node-log\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270356 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-netns\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270357 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-var-lib-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270517 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-cni-bin\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270556 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-config\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270588 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-script-lib\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270934 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-run-systemd\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.270990 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-slash\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271046 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-node-log\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271074 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-systemd-units\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271081 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-env-overrides\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271102 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-log-socket\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271109 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-etc-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271122 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-var-lib-openvswitch\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271134 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-kubelet\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271151 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23888411-56af-4cb6-871f-4804b5bcebf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271325 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-script-lib\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.271807 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23888411-56af-4cb6-871f-4804b5bcebf7-ovnkube-config\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.273559 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23888411-56af-4cb6-871f-4804b5bcebf7-ovn-node-metrics-cert\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.295719 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4876n\" (UniqueName: \"kubernetes.io/projected/23888411-56af-4cb6-871f-4804b5bcebf7-kube-api-access-4876n\") pod \"ovnkube-node-kzbqx\" (UID: \"23888411-56af-4cb6-871f-4804b5bcebf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.308114 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h5gln_60a84ebf-a188-4eee-8a9f-e02ac7adb8c8/kube-multus/2.log" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.308197 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h5gln" event={"ID":"60a84ebf-a188-4eee-8a9f-e02ac7adb8c8","Type":"ContainerStarted","Data":"c35ee1dc58fcea034b29deef8d2aab69c277b7bd13c6d6b75cd8bebd0280c9a2"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.313105 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.313463 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovnkube-controller/3.log" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.318252 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovn-acl-logging/0.log" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.318665 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qmqs_790411d3-0df3-4cc3-bf33-835dab4c2cb7/ovn-controller/0.log" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.318958 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.318991 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319001 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319012 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319022 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319058 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" exitCode=0 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319066 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" exitCode=143 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319074 4899 generic.go:334] "Generic (PLEG): container finished" podID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" exitCode=143 Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319098 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319129 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319145 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319157 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319170 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319183 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319197 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319228 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319236 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319242 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319249 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319258 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319265 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319272 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319278 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319289 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319301 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319309 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319317 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319325 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319332 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319339 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319346 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319353 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319360 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319367 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319377 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319388 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319396 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319403 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319410 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319419 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319426 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319433 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319442 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319449 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319455 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319465 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" event={"ID":"790411d3-0df3-4cc3-bf33-835dab4c2cb7","Type":"ContainerDied","Data":"ee61a426789003e32a0cde8bc9f972379128b2dd20aa8d64f515eb4b4120ed4b"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319487 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319496 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319503 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319509 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319516 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319523 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319529 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319535 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319542 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319550 4899 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319567 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.319743 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qmqs" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.363288 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.364340 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qmqs"] Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.369443 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qmqs"] Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.389644 4899 scope.go:117] "RemoveContainer" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.422658 4899 scope.go:117] "RemoveContainer" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.434898 4899 scope.go:117] "RemoveContainer" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.446575 4899 scope.go:117] "RemoveContainer" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.459333 4899 scope.go:117] "RemoveContainer" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.479575 4899 scope.go:117] "RemoveContainer" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.522117 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790411d3-0df3-4cc3-bf33-835dab4c2cb7" path="/var/lib/kubelet/pods/790411d3-0df3-4cc3-bf33-835dab4c2cb7/volumes" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.553423 4899 scope.go:117] "RemoveContainer" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.569166 4899 scope.go:117] "RemoveContainer" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.582976 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.583421 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.583475 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} err="failed to get container status \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.583510 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.583794 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": container with ID starting with 9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799 not found: ID does not exist" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.583823 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} err="failed to get container status \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": rpc error: code = NotFound desc = could not find container \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": container with ID starting with 9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.583847 4899 scope.go:117] "RemoveContainer" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.584059 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": container with ID starting with 0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d not found: ID does not exist" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584091 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} err="failed to get container status \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": rpc error: code = NotFound desc = could not find container \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": container with ID starting with 0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584113 4899 scope.go:117] "RemoveContainer" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.584597 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": container with ID starting with ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6 not found: ID does not exist" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584626 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} err="failed to get container status \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": rpc error: code = NotFound desc = could not find container \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": container with ID starting with ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584645 4899 scope.go:117] "RemoveContainer" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.584879 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": container with ID starting with a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292 not found: ID does not exist" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584909 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} err="failed to get container status \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": rpc error: code = NotFound desc = could not find container \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": container with ID starting with a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.584927 4899 scope.go:117] "RemoveContainer" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.585106 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": container with ID starting with 1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79 not found: ID does not exist" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585129 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} err="failed to get container status \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": rpc error: code = NotFound desc = could not find container \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": container with ID starting with 1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585144 4899 scope.go:117] "RemoveContainer" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.585346 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": container with ID starting with f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8 not found: ID does not exist" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585379 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} err="failed to get container status \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": rpc error: code = NotFound desc = could not find container \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": container with ID starting with f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585398 4899 scope.go:117] "RemoveContainer" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.585620 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": container with ID starting with c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64 not found: ID does not exist" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585646 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} err="failed to get container status \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": rpc error: code = NotFound desc = could not find container \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": container with ID starting with c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585663 4899 scope.go:117] "RemoveContainer" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.585897 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": container with ID starting with 645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41 not found: ID does not exist" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585917 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} err="failed to get container status \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": rpc error: code = NotFound desc = could not find container \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": container with ID starting with 645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.585928 4899 scope.go:117] "RemoveContainer" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: E0123 09:04:19.586130 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": container with ID starting with 01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80 not found: ID does not exist" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586153 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} err="failed to get container status \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": rpc error: code = NotFound desc = could not find container \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": container with ID starting with 01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586176 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586418 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} err="failed to get container status \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586464 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586840 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} err="failed to get container status \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": rpc error: code = NotFound desc = could not find container \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": container with ID starting with 9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.586872 4899 scope.go:117] "RemoveContainer" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587332 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} err="failed to get container status \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": rpc error: code = NotFound desc = could not find container \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": container with ID starting with 0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587388 4899 scope.go:117] "RemoveContainer" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587630 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} err="failed to get container status \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": rpc error: code = NotFound desc = could not find container \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": container with ID starting with ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587645 4899 scope.go:117] "RemoveContainer" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587862 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} err="failed to get container status \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": rpc error: code = NotFound desc = could not find container \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": container with ID starting with a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.587883 4899 scope.go:117] "RemoveContainer" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588094 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} err="failed to get container status \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": rpc error: code = NotFound desc = could not find container \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": container with ID starting with 1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588115 4899 scope.go:117] "RemoveContainer" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588400 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} err="failed to get container status \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": rpc error: code = NotFound desc = could not find container \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": container with ID starting with f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588421 4899 scope.go:117] "RemoveContainer" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588622 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} err="failed to get container status \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": rpc error: code = NotFound desc = could not find container \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": container with ID starting with c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588641 4899 scope.go:117] "RemoveContainer" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588851 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} err="failed to get container status \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": rpc error: code = NotFound desc = could not find container \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": container with ID starting with 645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.588876 4899 scope.go:117] "RemoveContainer" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589349 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} err="failed to get container status \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": rpc error: code = NotFound desc = could not find container \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": container with ID starting with 01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589377 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589600 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} err="failed to get container status \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589618 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589865 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} err="failed to get container status \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": rpc error: code = NotFound desc = could not find container \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": container with ID starting with 9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.589882 4899 scope.go:117] "RemoveContainer" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590081 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} err="failed to get container status \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": rpc error: code = NotFound desc = could not find container \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": container with ID starting with 0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590119 4899 scope.go:117] "RemoveContainer" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590397 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} err="failed to get container status \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": rpc error: code = NotFound desc = could not find container \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": container with ID starting with ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590418 4899 scope.go:117] "RemoveContainer" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590607 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} err="failed to get container status \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": rpc error: code = NotFound desc = could not find container \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": container with ID starting with a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590633 4899 scope.go:117] "RemoveContainer" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590876 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} err="failed to get container status \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": rpc error: code = NotFound desc = could not find container \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": container with ID starting with 1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.590899 4899 scope.go:117] "RemoveContainer" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591111 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} err="failed to get container status \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": rpc error: code = NotFound desc = could not find container \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": container with ID starting with f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591134 4899 scope.go:117] "RemoveContainer" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591379 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} err="failed to get container status \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": rpc error: code = NotFound desc = could not find container \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": container with ID starting with c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591395 4899 scope.go:117] "RemoveContainer" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591580 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} err="failed to get container status \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": rpc error: code = NotFound desc = could not find container \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": container with ID starting with 645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591599 4899 scope.go:117] "RemoveContainer" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591801 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} err="failed to get container status \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": rpc error: code = NotFound desc = could not find container \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": container with ID starting with 01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.591823 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592020 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} err="failed to get container status \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592037 4899 scope.go:117] "RemoveContainer" containerID="9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592228 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799"} err="failed to get container status \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": rpc error: code = NotFound desc = could not find container \"9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799\": container with ID starting with 9cb98af24cba8fdeae217e7d64823beee9656413f24829ade83160e2994e8799 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592254 4899 scope.go:117] "RemoveContainer" containerID="0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592581 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d"} err="failed to get container status \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": rpc error: code = NotFound desc = could not find container \"0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d\": container with ID starting with 0ec0c360f37a3c950942374b293e8679fd78f968eea76735af89b2bef4399b0d not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592600 4899 scope.go:117] "RemoveContainer" containerID="ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592792 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6"} err="failed to get container status \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": rpc error: code = NotFound desc = could not find container \"ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6\": container with ID starting with ce42e2077c026e21e816d3ae1f84173b6b4e6a0c5fb106314a443e3d87eeb7d6 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.592819 4899 scope.go:117] "RemoveContainer" containerID="a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593033 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292"} err="failed to get container status \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": rpc error: code = NotFound desc = could not find container \"a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292\": container with ID starting with a9875cc081a5c61fc8f2ce5864dbea81ed256531bf210c1a3bc6251e361db292 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593052 4899 scope.go:117] "RemoveContainer" containerID="1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593306 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79"} err="failed to get container status \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": rpc error: code = NotFound desc = could not find container \"1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79\": container with ID starting with 1a8d3eb604b0c5744bf151ab358b4072923d7998939292fa1a61cfa4321cae79 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593340 4899 scope.go:117] "RemoveContainer" containerID="f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593535 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8"} err="failed to get container status \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": rpc error: code = NotFound desc = could not find container \"f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8\": container with ID starting with f647c4f7ba20abfe57a8497db6b7f4f739a4d4ab7ebbc8f7d54eca7f4f3689e8 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593555 4899 scope.go:117] "RemoveContainer" containerID="c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593739 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64"} err="failed to get container status \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": rpc error: code = NotFound desc = could not find container \"c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64\": container with ID starting with c2e4f4a71d880645545067adda0c0b00ced6f90bd6f88b43510b2c06b4b83f64 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.593756 4899 scope.go:117] "RemoveContainer" containerID="645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.594023 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41"} err="failed to get container status \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": rpc error: code = NotFound desc = could not find container \"645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41\": container with ID starting with 645f3a8976fe4b5add6b00f0baa1650d75a8e2ce21dc0ebb2a5dbf7a22cbfc41 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.594048 4899 scope.go:117] "RemoveContainer" containerID="01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.594304 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80"} err="failed to get container status \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": rpc error: code = NotFound desc = could not find container \"01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80\": container with ID starting with 01e182a6cc19d45c5049a01e7dcbe93f9225100b06bc7753d4dfaa1ad4999b80 not found: ID does not exist" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.594328 4899 scope.go:117] "RemoveContainer" containerID="ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2" Jan 23 09:04:19 crc kubenswrapper[4899]: I0123 09:04:19.594559 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2"} err="failed to get container status \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": rpc error: code = NotFound desc = could not find container \"ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2\": container with ID starting with ec6995fdc0c16c6de64a0e429cd3c4f7132dcf71a42a0e853533d20ae65b89b2 not found: ID does not exist" Jan 23 09:04:20 crc kubenswrapper[4899]: I0123 09:04:20.341471 4899 generic.go:334] "Generic (PLEG): container finished" podID="23888411-56af-4cb6-871f-4804b5bcebf7" containerID="ac27cb73ce272a193984ed52c6a10fd498acb99d820d46d08020d6b700418a1c" exitCode=0 Jan 23 09:04:20 crc kubenswrapper[4899]: I0123 09:04:20.341506 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerDied","Data":"ac27cb73ce272a193984ed52c6a10fd498acb99d820d46d08020d6b700418a1c"} Jan 23 09:04:20 crc kubenswrapper[4899]: I0123 09:04:20.341704 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"900d70c4ad4566694da05f299a70b3aafe294c2462823d52cad0b61902eab3a5"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.349939 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"219a6eaa316f5a2fe4dc42726f8c8f409cf8db601b3c25d1d94ba27134a1f2c7"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.350247 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"e7565309489e9076ed9ab3af62127a1dae70d2aced0ae0a6e959da6f3c401565"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.350258 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"d1ce036d386a77b414bd04533cc9c5b19dab5405abbc8249d1855df2644bf232"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.350266 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"3d8a13d57b400d05af8501679323db5121a74255dfde94638b0bc99bc0d82b6d"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.350275 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"921036fe779831af8204aa41bb99bee318be746be012f74f4d15f7b499b2ed34"} Jan 23 09:04:21 crc kubenswrapper[4899]: I0123 09:04:21.350284 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"f0441864e4d30af0d02bbc949bcd2d89fa3ab798387fab7d8a13f37d99b8f0bb"} Jan 23 09:04:23 crc kubenswrapper[4899]: I0123 09:04:23.364137 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"961d07afc2635c89aef8042462918d2bf4ec6d53e8d829e72da8b4f8eb9125d8"} Jan 23 09:04:26 crc kubenswrapper[4899]: I0123 09:04:26.383307 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" event={"ID":"23888411-56af-4cb6-871f-4804b5bcebf7","Type":"ContainerStarted","Data":"960cfb4c2f02417fb783e56df04a0667eaeb8a920cdd9093b70e9861b1d9df89"} Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.137998 4899 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.387442 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.387734 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.387753 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.422010 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" podStartSLOduration=9.421991606 podStartE2EDuration="9.421991606s" podCreationTimestamp="2026-01-23 09:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 09:04:27.417143013 +0000 UTC m=+754.284113400" watchObservedRunningTime="2026-01-23 09:04:27.421991606 +0000 UTC m=+754.288961993" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.423733 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:27 crc kubenswrapper[4899]: I0123 09:04:27.427912 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:31 crc kubenswrapper[4899]: I0123 09:04:31.685679 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:04:31 crc kubenswrapper[4899]: I0123 09:04:31.685998 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:04:48 crc kubenswrapper[4899]: I0123 09:04:48.965739 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:04:48 crc kubenswrapper[4899]: I0123 09:04:48.971067 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:48 crc kubenswrapper[4899]: I0123 09:04:48.977945 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.043934 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.044042 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr995\" (UniqueName: \"kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.044093 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.145707 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.145758 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.145811 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr995\" (UniqueName: \"kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.146305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.146432 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.171794 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr995\" (UniqueName: \"kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995\") pod \"redhat-operators-kkllx\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.289888 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.336542 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kzbqx" Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.534023 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:04:49 crc kubenswrapper[4899]: I0123 09:04:49.589112 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerStarted","Data":"4fbe37848b6f33da7f4252e39dbacc34351c892add005f5331975fe448898ac8"} Jan 23 09:04:50 crc kubenswrapper[4899]: I0123 09:04:50.596049 4899 generic.go:334] "Generic (PLEG): container finished" podID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerID="14913001f1e09ad3d167d2df62d7991487670854880638f88ef197e69281d4d2" exitCode=0 Jan 23 09:04:50 crc kubenswrapper[4899]: I0123 09:04:50.596107 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerDied","Data":"14913001f1e09ad3d167d2df62d7991487670854880638f88ef197e69281d4d2"} Jan 23 09:04:50 crc kubenswrapper[4899]: I0123 09:04:50.598182 4899 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 23 09:04:51 crc kubenswrapper[4899]: I0123 09:04:51.603619 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerStarted","Data":"55424742485fe46dfbfae55a032461d68ff08b61b9adfdd805f527d591fd94a6"} Jan 23 09:04:52 crc kubenswrapper[4899]: I0123 09:04:52.610129 4899 generic.go:334] "Generic (PLEG): container finished" podID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerID="55424742485fe46dfbfae55a032461d68ff08b61b9adfdd805f527d591fd94a6" exitCode=0 Jan 23 09:04:52 crc kubenswrapper[4899]: I0123 09:04:52.610284 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerDied","Data":"55424742485fe46dfbfae55a032461d68ff08b61b9adfdd805f527d591fd94a6"} Jan 23 09:04:53 crc kubenswrapper[4899]: I0123 09:04:53.621289 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerStarted","Data":"99694eba2dbc9c1e706d76d123c169fd73430750215b7bed57036a439c4daeb1"} Jan 23 09:04:53 crc kubenswrapper[4899]: I0123 09:04:53.642003 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kkllx" podStartSLOduration=3.232045554 podStartE2EDuration="5.64198747s" podCreationTimestamp="2026-01-23 09:04:48 +0000 UTC" firstStartedPulling="2026-01-23 09:04:50.597910567 +0000 UTC m=+777.464880954" lastFinishedPulling="2026-01-23 09:04:53.007852483 +0000 UTC m=+779.874822870" observedRunningTime="2026-01-23 09:04:53.638050452 +0000 UTC m=+780.505020859" watchObservedRunningTime="2026-01-23 09:04:53.64198747 +0000 UTC m=+780.508957857" Jan 23 09:04:59 crc kubenswrapper[4899]: I0123 09:04:59.290057 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:59 crc kubenswrapper[4899]: I0123 09:04:59.290336 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:59 crc kubenswrapper[4899]: I0123 09:04:59.352909 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:59 crc kubenswrapper[4899]: I0123 09:04:59.688170 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:04:59 crc kubenswrapper[4899]: I0123 09:04:59.732714 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.662432 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kkllx" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="registry-server" containerID="cri-o://99694eba2dbc9c1e706d76d123c169fd73430750215b7bed57036a439c4daeb1" gracePeriod=2 Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.686158 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.686266 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.686326 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.687103 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:05:01 crc kubenswrapper[4899]: I0123 09:05:01.687187 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4" gracePeriod=600 Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.685338 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerDied","Data":"99694eba2dbc9c1e706d76d123c169fd73430750215b7bed57036a439c4daeb1"} Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.685202 4899 generic.go:334] "Generic (PLEG): container finished" podID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerID="99694eba2dbc9c1e706d76d123c169fd73430750215b7bed57036a439c4daeb1" exitCode=0 Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.692390 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4" exitCode=0 Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.692443 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4"} Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.692484 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41"} Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.692506 4899 scope.go:117] "RemoveContainer" containerID="c9a13017cbd8fa75c3ce0f1375e22ced93e304a63a6dab169c786eeabe8a8ae9" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.739281 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.850463 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content\") pod \"b329ea65-b54b-4a88-b6f3-76c8fab02654\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.850534 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr995\" (UniqueName: \"kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995\") pod \"b329ea65-b54b-4a88-b6f3-76c8fab02654\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.850564 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities\") pod \"b329ea65-b54b-4a88-b6f3-76c8fab02654\" (UID: \"b329ea65-b54b-4a88-b6f3-76c8fab02654\") " Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.852223 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities" (OuterVolumeSpecName: "utilities") pod "b329ea65-b54b-4a88-b6f3-76c8fab02654" (UID: "b329ea65-b54b-4a88-b6f3-76c8fab02654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.855546 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995" (OuterVolumeSpecName: "kube-api-access-pr995") pod "b329ea65-b54b-4a88-b6f3-76c8fab02654" (UID: "b329ea65-b54b-4a88-b6f3-76c8fab02654"). InnerVolumeSpecName "kube-api-access-pr995". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.951974 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr995\" (UniqueName: \"kubernetes.io/projected/b329ea65-b54b-4a88-b6f3-76c8fab02654-kube-api-access-pr995\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.952002 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:04 crc kubenswrapper[4899]: I0123 09:05:04.963437 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b329ea65-b54b-4a88-b6f3-76c8fab02654" (UID: "b329ea65-b54b-4a88-b6f3-76c8fab02654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.053383 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b329ea65-b54b-4a88-b6f3-76c8fab02654-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.705488 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkllx" event={"ID":"b329ea65-b54b-4a88-b6f3-76c8fab02654","Type":"ContainerDied","Data":"4fbe37848b6f33da7f4252e39dbacc34351c892add005f5331975fe448898ac8"} Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.705548 4899 scope.go:117] "RemoveContainer" containerID="99694eba2dbc9c1e706d76d123c169fd73430750215b7bed57036a439c4daeb1" Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.705701 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkllx" Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.728833 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.732690 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kkllx"] Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.737759 4899 scope.go:117] "RemoveContainer" containerID="55424742485fe46dfbfae55a032461d68ff08b61b9adfdd805f527d591fd94a6" Jan 23 09:05:05 crc kubenswrapper[4899]: I0123 09:05:05.757141 4899 scope.go:117] "RemoveContainer" containerID="14913001f1e09ad3d167d2df62d7991487670854880638f88ef197e69281d4d2" Jan 23 09:05:07 crc kubenswrapper[4899]: I0123 09:05:07.522609 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" path="/var/lib/kubelet/pods/b329ea65-b54b-4a88-b6f3-76c8fab02654/volumes" Jan 23 09:05:46 crc kubenswrapper[4899]: I0123 09:05:46.672407 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 09:05:46 crc kubenswrapper[4899]: I0123 09:05:46.673282 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4x76s" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="registry-server" containerID="cri-o://6fb0fa33f1e78ae4eea1410da40a74a18ab6cb1e9b1831dabc3d616ee581091e" gracePeriod=30 Jan 23 09:05:46 crc kubenswrapper[4899]: I0123 09:05:46.959504 4899 generic.go:334] "Generic (PLEG): container finished" podID="1d620e98-4c84-4ff3-a311-c839033909aa" containerID="6fb0fa33f1e78ae4eea1410da40a74a18ab6cb1e9b1831dabc3d616ee581091e" exitCode=0 Jan 23 09:05:46 crc kubenswrapper[4899]: I0123 09:05:46.959561 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerDied","Data":"6fb0fa33f1e78ae4eea1410da40a74a18ab6cb1e9b1831dabc3d616ee581091e"} Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.513987 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.599172 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities\") pod \"1d620e98-4c84-4ff3-a311-c839033909aa\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.599242 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrzbn\" (UniqueName: \"kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn\") pod \"1d620e98-4c84-4ff3-a311-c839033909aa\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.599269 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content\") pod \"1d620e98-4c84-4ff3-a311-c839033909aa\" (UID: \"1d620e98-4c84-4ff3-a311-c839033909aa\") " Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.600621 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities" (OuterVolumeSpecName: "utilities") pod "1d620e98-4c84-4ff3-a311-c839033909aa" (UID: "1d620e98-4c84-4ff3-a311-c839033909aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.607375 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn" (OuterVolumeSpecName: "kube-api-access-jrzbn") pod "1d620e98-4c84-4ff3-a311-c839033909aa" (UID: "1d620e98-4c84-4ff3-a311-c839033909aa"). InnerVolumeSpecName "kube-api-access-jrzbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.625109 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d620e98-4c84-4ff3-a311-c839033909aa" (UID: "1d620e98-4c84-4ff3-a311-c839033909aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.700667 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.700703 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrzbn\" (UniqueName: \"kubernetes.io/projected/1d620e98-4c84-4ff3-a311-c839033909aa-kube-api-access-jrzbn\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.700717 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d620e98-4c84-4ff3-a311-c839033909aa-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.966389 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4x76s" event={"ID":"1d620e98-4c84-4ff3-a311-c839033909aa","Type":"ContainerDied","Data":"56a5c3ddccbde34946c8a8b7bb764cde9b4d16b397839716a356df486df5127d"} Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.966963 4899 scope.go:117] "RemoveContainer" containerID="6fb0fa33f1e78ae4eea1410da40a74a18ab6cb1e9b1831dabc3d616ee581091e" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.966922 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4x76s" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.985409 4899 scope.go:117] "RemoveContainer" containerID="f98bac878649c293ca9f762a7f3256ec30005e4f09d0270f648749d8e192b7ec" Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.994949 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 09:05:47 crc kubenswrapper[4899]: I0123 09:05:47.999320 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4x76s"] Jan 23 09:05:48 crc kubenswrapper[4899]: I0123 09:05:48.021807 4899 scope.go:117] "RemoveContainer" containerID="b804b2a710ef73a83d124ade32ff344d36a739d0bcf4182c3f3df2d1c3915cae" Jan 23 09:05:49 crc kubenswrapper[4899]: I0123 09:05:49.520995 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" path="/var/lib/kubelet/pods/1d620e98-4c84-4ff3-a311-c839033909aa/volumes" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.663863 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch"] Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664079 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="extract-utilities" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664091 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="extract-utilities" Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664103 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664111 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664122 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664131 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664142 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="extract-content" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664150 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="extract-content" Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664160 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="extract-utilities" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664167 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="extract-utilities" Jan 23 09:05:51 crc kubenswrapper[4899]: E0123 09:05:51.664176 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="extract-content" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664183 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="extract-content" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664324 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d620e98-4c84-4ff3-a311-c839033909aa" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.664344 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="b329ea65-b54b-4a88-b6f3-76c8fab02654" containerName="registry-server" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.665181 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.667684 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.685282 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch"] Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.747417 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.747488 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k84kn\" (UniqueName: \"kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.747570 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.848852 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.848951 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k84kn\" (UniqueName: \"kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.849055 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.849532 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.849636 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.884492 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k84kn\" (UniqueName: \"kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:51 crc kubenswrapper[4899]: I0123 09:05:51.978530 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:52 crc kubenswrapper[4899]: I0123 09:05:52.186564 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch"] Jan 23 09:05:52 crc kubenswrapper[4899]: I0123 09:05:52.995468 4899 generic.go:334] "Generic (PLEG): container finished" podID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerID="69591d39ddd89ecc063d0783a2876f3760d0321485b99dc39602a7e55ffcdd34" exitCode=0 Jan 23 09:05:52 crc kubenswrapper[4899]: I0123 09:05:52.995610 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerDied","Data":"69591d39ddd89ecc063d0783a2876f3760d0321485b99dc39602a7e55ffcdd34"} Jan 23 09:05:52 crc kubenswrapper[4899]: I0123 09:05:52.995827 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerStarted","Data":"eb7798afc71b715d469de916463662cee7f3fee6d0bf209f4fd234ff0672dcb9"} Jan 23 09:05:55 crc kubenswrapper[4899]: I0123 09:05:55.007142 4899 generic.go:334] "Generic (PLEG): container finished" podID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerID="04a0bfa96dc67017318a40ad5c7d5de52aec6a2367be9a1b3ee55dec63113c02" exitCode=0 Jan 23 09:05:55 crc kubenswrapper[4899]: I0123 09:05:55.007229 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerDied","Data":"04a0bfa96dc67017318a40ad5c7d5de52aec6a2367be9a1b3ee55dec63113c02"} Jan 23 09:05:56 crc kubenswrapper[4899]: I0123 09:05:56.016299 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerStarted","Data":"617c16de70c16ede7a290a7e44604d690c66fe7634281c225f075a852c4c3ff2"} Jan 23 09:05:56 crc kubenswrapper[4899]: I0123 09:05:56.046205 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" podStartSLOduration=4.006773963 podStartE2EDuration="5.046181713s" podCreationTimestamp="2026-01-23 09:05:51 +0000 UTC" firstStartedPulling="2026-01-23 09:05:52.996989523 +0000 UTC m=+839.863959930" lastFinishedPulling="2026-01-23 09:05:54.036397293 +0000 UTC m=+840.903367680" observedRunningTime="2026-01-23 09:05:56.042174965 +0000 UTC m=+842.909145392" watchObservedRunningTime="2026-01-23 09:05:56.046181713 +0000 UTC m=+842.913152140" Jan 23 09:05:57 crc kubenswrapper[4899]: I0123 09:05:57.023838 4899 generic.go:334] "Generic (PLEG): container finished" podID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerID="617c16de70c16ede7a290a7e44604d690c66fe7634281c225f075a852c4c3ff2" exitCode=0 Jan 23 09:05:57 crc kubenswrapper[4899]: I0123 09:05:57.023876 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerDied","Data":"617c16de70c16ede7a290a7e44604d690c66fe7634281c225f075a852c4c3ff2"} Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.256789 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.325508 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle\") pod \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.325719 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k84kn\" (UniqueName: \"kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn\") pod \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.325759 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util\") pod \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\" (UID: \"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a\") " Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.328414 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle" (OuterVolumeSpecName: "bundle") pod "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" (UID: "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.331331 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn" (OuterVolumeSpecName: "kube-api-access-k84kn") pod "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" (UID: "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a"). InnerVolumeSpecName "kube-api-access-k84kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.336989 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util" (OuterVolumeSpecName: "util") pod "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" (UID: "6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.426854 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.426909 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:58 crc kubenswrapper[4899]: I0123 09:05:58.426922 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k84kn\" (UniqueName: \"kubernetes.io/projected/6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a-kube-api-access-k84kn\") on node \"crc\" DevicePath \"\"" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.036906 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" event={"ID":"6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a","Type":"ContainerDied","Data":"eb7798afc71b715d469de916463662cee7f3fee6d0bf209f4fd234ff0672dcb9"} Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.036946 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb7798afc71b715d469de916463662cee7f3fee6d0bf209f4fd234ff0672dcb9" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.037009 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.860176 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5"] Jan 23 09:05:59 crc kubenswrapper[4899]: E0123 09:05:59.860410 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="util" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.860422 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="util" Jan 23 09:05:59 crc kubenswrapper[4899]: E0123 09:05:59.860438 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="extract" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.860460 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="extract" Jan 23 09:05:59 crc kubenswrapper[4899]: E0123 09:05:59.860478 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="pull" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.860484 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="pull" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.860562 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a" containerName="extract" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.861224 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.864438 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.866322 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5"] Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.946297 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.946440 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw829\" (UniqueName: \"kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:05:59 crc kubenswrapper[4899]: I0123 09:05:59.946495 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.047530 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.048074 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw829\" (UniqueName: \"kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.048101 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.048254 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.048639 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.071350 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw829\" (UniqueName: \"kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.180438 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.350072 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5"] Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.878184 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t"] Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.879711 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.887360 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t"] Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.960713 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spwf2\" (UniqueName: \"kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.960774 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:00 crc kubenswrapper[4899]: I0123 09:06:00.960875 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.045687 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" event={"ID":"327637d2-083d-4465-90a4-be769c435c9f","Type":"ContainerStarted","Data":"9630c9a7e97d500fe1280806c395681c44b373b3e0fbb67ec9b692eddc2e2b2e"} Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.061792 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.061845 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spwf2\" (UniqueName: \"kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.061873 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.062328 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.062404 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.078550 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spwf2\" (UniqueName: \"kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.232852 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:01 crc kubenswrapper[4899]: I0123 09:06:01.417612 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t"] Jan 23 09:06:02 crc kubenswrapper[4899]: I0123 09:06:02.058589 4899 generic.go:334] "Generic (PLEG): container finished" podID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerID="2f07bb39697cf16188b03b4456388f6ccaf84ae41114562ede85e9ca02b953e5" exitCode=0 Jan 23 09:06:02 crc kubenswrapper[4899]: I0123 09:06:02.058707 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerDied","Data":"2f07bb39697cf16188b03b4456388f6ccaf84ae41114562ede85e9ca02b953e5"} Jan 23 09:06:02 crc kubenswrapper[4899]: I0123 09:06:02.058807 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerStarted","Data":"3dfc9a8e43b3f0de4d1f33a5488a825cba37ec17a75fc7df30c3aa2b633b6862"} Jan 23 09:06:02 crc kubenswrapper[4899]: I0123 09:06:02.074633 4899 generic.go:334] "Generic (PLEG): container finished" podID="327637d2-083d-4465-90a4-be769c435c9f" containerID="224f9ef68ddf3c0b5f93b45150764da074eb47ffb5cc09c335f1abe8d030a24a" exitCode=0 Jan 23 09:06:02 crc kubenswrapper[4899]: I0123 09:06:02.074694 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" event={"ID":"327637d2-083d-4465-90a4-be769c435c9f","Type":"ContainerDied","Data":"224f9ef68ddf3c0b5f93b45150764da074eb47ffb5cc09c335f1abe8d030a24a"} Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.084539 4899 generic.go:334] "Generic (PLEG): container finished" podID="327637d2-083d-4465-90a4-be769c435c9f" containerID="7e8c10169b5c36de7df3ca23e5a16c285ee40050104459de9bce98a12590a907" exitCode=0 Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.084610 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" event={"ID":"327637d2-083d-4465-90a4-be769c435c9f","Type":"ContainerDied","Data":"7e8c10169b5c36de7df3ca23e5a16c285ee40050104459de9bce98a12590a907"} Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.456318 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-44llv"] Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.457801 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.475684 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44llv"] Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.616681 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-catalog-content\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.616729 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-utilities\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.616752 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xtg2\" (UniqueName: \"kubernetes.io/projected/58968c93-e09f-4d90-acf8-4bb499beb6e5-kube-api-access-2xtg2\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.726967 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-catalog-content\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.727047 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-utilities\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.727073 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xtg2\" (UniqueName: \"kubernetes.io/projected/58968c93-e09f-4d90-acf8-4bb499beb6e5-kube-api-access-2xtg2\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.727916 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-catalog-content\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.728162 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58968c93-e09f-4d90-acf8-4bb499beb6e5-utilities\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:04 crc kubenswrapper[4899]: I0123 09:06:04.824390 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xtg2\" (UniqueName: \"kubernetes.io/projected/58968c93-e09f-4d90-acf8-4bb499beb6e5-kube-api-access-2xtg2\") pod \"certified-operators-44llv\" (UID: \"58968c93-e09f-4d90-acf8-4bb499beb6e5\") " pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:05 crc kubenswrapper[4899]: I0123 09:06:05.081883 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:05 crc kubenswrapper[4899]: I0123 09:06:05.091711 4899 generic.go:334] "Generic (PLEG): container finished" podID="327637d2-083d-4465-90a4-be769c435c9f" containerID="6c8497dd4dd1954d62ffdab0eed3c162c989617f9c6fb6974081dfab33efc923" exitCode=0 Jan 23 09:06:05 crc kubenswrapper[4899]: I0123 09:06:05.091760 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" event={"ID":"327637d2-083d-4465-90a4-be769c435c9f","Type":"ContainerDied","Data":"6c8497dd4dd1954d62ffdab0eed3c162c989617f9c6fb6974081dfab33efc923"} Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.739407 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.813792 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util\") pod \"327637d2-083d-4465-90a4-be769c435c9f\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.814361 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw829\" (UniqueName: \"kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829\") pod \"327637d2-083d-4465-90a4-be769c435c9f\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.814400 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle\") pod \"327637d2-083d-4465-90a4-be769c435c9f\" (UID: \"327637d2-083d-4465-90a4-be769c435c9f\") " Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.815293 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle" (OuterVolumeSpecName: "bundle") pod "327637d2-083d-4465-90a4-be769c435c9f" (UID: "327637d2-083d-4465-90a4-be769c435c9f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.815959 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.827845 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829" (OuterVolumeSpecName: "kube-api-access-nw829") pod "327637d2-083d-4465-90a4-be769c435c9f" (UID: "327637d2-083d-4465-90a4-be769c435c9f"). InnerVolumeSpecName "kube-api-access-nw829". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.838398 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util" (OuterVolumeSpecName: "util") pod "327637d2-083d-4465-90a4-be769c435c9f" (UID: "327637d2-083d-4465-90a4-be769c435c9f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.910471 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7"] Jan 23 09:06:10 crc kubenswrapper[4899]: E0123 09:06:10.910724 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="util" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.910740 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="util" Jan 23 09:06:10 crc kubenswrapper[4899]: E0123 09:06:10.910751 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="pull" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.910757 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="pull" Jan 23 09:06:10 crc kubenswrapper[4899]: E0123 09:06:10.910772 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="extract" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.910778 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="extract" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.910881 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="327637d2-083d-4465-90a4-be769c435c9f" containerName="extract" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.911321 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.915617 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.916057 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-hprc6" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.916291 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.921869 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw829\" (UniqueName: \"kubernetes.io/projected/327637d2-083d-4465-90a4-be769c435c9f-kube-api-access-nw829\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.921906 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/327637d2-083d-4465-90a4-be769c435c9f-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:10 crc kubenswrapper[4899]: I0123 09:06:10.926290 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.022864 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb57g\" (UniqueName: \"kubernetes.io/projected/a9bf5caa-85e0-4b70-8145-d6c25fa15999-kube-api-access-lb57g\") pod \"obo-prometheus-operator-68bc856cb9-5dpg7\" (UID: \"a9bf5caa-85e0-4b70-8145-d6c25fa15999\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.083265 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.083940 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.085824 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.085863 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-kkk4r" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.088669 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.089459 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.106250 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.106295 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.123947 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb57g\" (UniqueName: \"kubernetes.io/projected/a9bf5caa-85e0-4b70-8145-d6c25fa15999-kube-api-access-lb57g\") pod \"obo-prometheus-operator-68bc856cb9-5dpg7\" (UID: \"a9bf5caa-85e0-4b70-8145-d6c25fa15999\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.152754 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb57g\" (UniqueName: \"kubernetes.io/projected/a9bf5caa-85e0-4b70-8145-d6c25fa15999-kube-api-access-lb57g\") pod \"obo-prometheus-operator-68bc856cb9-5dpg7\" (UID: \"a9bf5caa-85e0-4b70-8145-d6c25fa15999\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.218447 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-g9rtj"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.219711 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.221502 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerStarted","Data":"795a9d5818d46cefef0a41671fe91e98ac7cc216fc4718f0f8133bf8315fbd12"} Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.221782 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-59hwk" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.224117 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.224583 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.224617 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.224643 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.224698 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.230683 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-g9rtj"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.231727 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" event={"ID":"327637d2-083d-4465-90a4-be769c435c9f","Type":"ContainerDied","Data":"9630c9a7e97d500fe1280806c395681c44b373b3e0fbb67ec9b692eddc2e2b2e"} Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.231789 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9630c9a7e97d500fe1280806c395681c44b373b3e0fbb67ec9b692eddc2e2b2e" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.231866 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.235193 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.298954 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44llv"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.325908 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.325965 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.326005 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.326082 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7pn\" (UniqueName: \"kubernetes.io/projected/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-kube-api-access-md7pn\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.326128 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-observability-operator-tls\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.326169 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.331014 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.331682 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e90996a1-43d9-44c4-b23e-e8b110764326-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw\" (UID: \"e90996a1-43d9-44c4-b23e-e8b110764326\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.331699 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.332138 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab9f7afb-bc13-43ce-b218-bdb18a103b51-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd\" (UID: \"ab9f7afb-bc13-43ce-b218-bdb18a103b51\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.427059 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7pn\" (UniqueName: \"kubernetes.io/projected/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-kube-api-access-md7pn\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.427108 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-observability-operator-tls\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.430882 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-observability-operator-tls\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.441782 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.454804 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7pn\" (UniqueName: \"kubernetes.io/projected/6a152f52-cbf3-43d4-b9fb-0b312c9c0737-kube-api-access-md7pn\") pod \"observability-operator-59bdc8b94-g9rtj\" (UID: \"6a152f52-cbf3-43d4-b9fb-0b312c9c0737\") " pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.461034 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-vgmkw"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.462607 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.465664 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-5867h" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.471972 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.474312 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.527826 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw78x\" (UniqueName: \"kubernetes.io/projected/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-kube-api-access-nw78x\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.527866 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-openshift-service-ca\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.549550 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.629972 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw78x\" (UniqueName: \"kubernetes.io/projected/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-kube-api-access-nw78x\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.630337 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-openshift-service-ca\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.631444 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-openshift-service-ca\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.658284 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw78x\" (UniqueName: \"kubernetes.io/projected/28fbc9c0-2fdd-46d3-bc64-bcec54804e17-kube-api-access-nw78x\") pod \"perses-operator-5bf474d74f-vgmkw\" (UID: \"28fbc9c0-2fdd-46d3-bc64-bcec54804e17\") " pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.752639 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw"] Jan 23 09:06:11 crc kubenswrapper[4899]: W0123 09:06:11.773537 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode90996a1_43d9_44c4_b23e_e8b110764326.slice/crio-f65a1b35ab67922ddccfc12fcb99287ae199e95d1bb7f966120c1cb03eb06994 WatchSource:0}: Error finding container f65a1b35ab67922ddccfc12fcb99287ae199e95d1bb7f966120c1cb03eb06994: Status 404 returned error can't find the container with id f65a1b35ab67922ddccfc12fcb99287ae199e95d1bb7f966120c1cb03eb06994 Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.791778 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.837107 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-vgmkw"] Jan 23 09:06:11 crc kubenswrapper[4899]: I0123 09:06:11.891856 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd"] Jan 23 09:06:11 crc kubenswrapper[4899]: W0123 09:06:11.995193 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab9f7afb_bc13_43ce_b218_bdb18a103b51.slice/crio-9a15b1821d715d5c2f9995cb166a6f164d3f51fd2701fd14bf4b9b0e8e45ec5f WatchSource:0}: Error finding container 9a15b1821d715d5c2f9995cb166a6f164d3f51fd2701fd14bf4b9b0e8e45ec5f: Status 404 returned error can't find the container with id 9a15b1821d715d5c2f9995cb166a6f164d3f51fd2701fd14bf4b9b0e8e45ec5f Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.017199 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-g9rtj"] Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.111011 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-vgmkw"] Jan 23 09:06:12 crc kubenswrapper[4899]: W0123 09:06:12.115935 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28fbc9c0_2fdd_46d3_bc64_bcec54804e17.slice/crio-64b41867f118759d65733e5168437e70a8c144b4832fc8b71b4703343eff1958 WatchSource:0}: Error finding container 64b41867f118759d65733e5168437e70a8c144b4832fc8b71b4703343eff1958: Status 404 returned error can't find the container with id 64b41867f118759d65733e5168437e70a8c144b4832fc8b71b4703343eff1958 Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.238286 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" event={"ID":"6a152f52-cbf3-43d4-b9fb-0b312c9c0737","Type":"ContainerStarted","Data":"108d2c1c4c9d3ad87600e3636d442412ecb458734c07778bef54a601141a6490"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.239726 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" event={"ID":"ab9f7afb-bc13-43ce-b218-bdb18a103b51","Type":"ContainerStarted","Data":"9a15b1821d715d5c2f9995cb166a6f164d3f51fd2701fd14bf4b9b0e8e45ec5f"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.241075 4899 generic.go:334] "Generic (PLEG): container finished" podID="58968c93-e09f-4d90-acf8-4bb499beb6e5" containerID="225689ceb2d3f8ab0328d64bf3c8aeddb76495ca2465eb55d2b4e8d7d0e1fba6" exitCode=0 Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.241135 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44llv" event={"ID":"58968c93-e09f-4d90-acf8-4bb499beb6e5","Type":"ContainerDied","Data":"225689ceb2d3f8ab0328d64bf3c8aeddb76495ca2465eb55d2b4e8d7d0e1fba6"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.241159 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44llv" event={"ID":"58968c93-e09f-4d90-acf8-4bb499beb6e5","Type":"ContainerStarted","Data":"b5dd2001a9cfb917fc3b73b34e817b5f00097cb2ea7a430282b893d9b944f666"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.243463 4899 generic.go:334] "Generic (PLEG): container finished" podID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerID="795a9d5818d46cefef0a41671fe91e98ac7cc216fc4718f0f8133bf8315fbd12" exitCode=0 Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.243514 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerDied","Data":"795a9d5818d46cefef0a41671fe91e98ac7cc216fc4718f0f8133bf8315fbd12"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.247240 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" event={"ID":"e90996a1-43d9-44c4-b23e-e8b110764326","Type":"ContainerStarted","Data":"f65a1b35ab67922ddccfc12fcb99287ae199e95d1bb7f966120c1cb03eb06994"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.249568 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" event={"ID":"28fbc9c0-2fdd-46d3-bc64-bcec54804e17","Type":"ContainerStarted","Data":"64b41867f118759d65733e5168437e70a8c144b4832fc8b71b4703343eff1958"} Jan 23 09:06:12 crc kubenswrapper[4899]: I0123 09:06:12.250589 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" event={"ID":"a9bf5caa-85e0-4b70-8145-d6c25fa15999","Type":"ContainerStarted","Data":"976dd4c4abb1b75cbd9b7e7799755939854c7e7ca8b7e30ad8bfa923c586262b"} Jan 23 09:06:13 crc kubenswrapper[4899]: I0123 09:06:13.269649 4899 generic.go:334] "Generic (PLEG): container finished" podID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerID="ed5052de8a735a216d81150d4fddd6b840d02d6b30b2fcc5ce9f6653605d7fad" exitCode=0 Jan 23 09:06:13 crc kubenswrapper[4899]: I0123 09:06:13.269843 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerDied","Data":"ed5052de8a735a216d81150d4fddd6b840d02d6b30b2fcc5ce9f6653605d7fad"} Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.670695 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.797189 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spwf2\" (UniqueName: \"kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2\") pod \"d8537859-a9d3-426f-9f67-631b7b5763fc\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.797349 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle\") pod \"d8537859-a9d3-426f-9f67-631b7b5763fc\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.798385 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle" (OuterVolumeSpecName: "bundle") pod "d8537859-a9d3-426f-9f67-631b7b5763fc" (UID: "d8537859-a9d3-426f-9f67-631b7b5763fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.798469 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util\") pod \"d8537859-a9d3-426f-9f67-631b7b5763fc\" (UID: \"d8537859-a9d3-426f-9f67-631b7b5763fc\") " Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.799720 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.810687 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util" (OuterVolumeSpecName: "util") pod "d8537859-a9d3-426f-9f67-631b7b5763fc" (UID: "d8537859-a9d3-426f-9f67-631b7b5763fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.819582 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2" (OuterVolumeSpecName: "kube-api-access-spwf2") pod "d8537859-a9d3-426f-9f67-631b7b5763fc" (UID: "d8537859-a9d3-426f-9f67-631b7b5763fc"). InnerVolumeSpecName "kube-api-access-spwf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.901103 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spwf2\" (UniqueName: \"kubernetes.io/projected/d8537859-a9d3-426f-9f67-631b7b5763fc-kube-api-access-spwf2\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:14 crc kubenswrapper[4899]: I0123 09:06:14.901142 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8537859-a9d3-426f-9f67-631b7b5763fc-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:06:15 crc kubenswrapper[4899]: I0123 09:06:15.298220 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" event={"ID":"d8537859-a9d3-426f-9f67-631b7b5763fc","Type":"ContainerDied","Data":"3dfc9a8e43b3f0de4d1f33a5488a825cba37ec17a75fc7df30c3aa2b633b6862"} Jan 23 09:06:15 crc kubenswrapper[4899]: I0123 09:06:15.298258 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dfc9a8e43b3f0de4d1f33a5488a825cba37ec17a75fc7df30c3aa2b633b6862" Jan 23 09:06:15 crc kubenswrapper[4899]: I0123 09:06:15.298318 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.601824 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-7997b94565-8v5lj"] Jan 23 09:06:20 crc kubenswrapper[4899]: E0123 09:06:20.602611 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="util" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.602627 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="util" Jan 23 09:06:20 crc kubenswrapper[4899]: E0123 09:06:20.602640 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="pull" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.602647 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="pull" Jan 23 09:06:20 crc kubenswrapper[4899]: E0123 09:06:20.602662 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="extract" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.602670 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="extract" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.602782 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8537859-a9d3-426f-9f67-631b7b5763fc" containerName="extract" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.603270 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.606423 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.607185 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-6nf7l" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.607585 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.608039 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.625368 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7997b94565-8v5lj"] Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.760642 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-apiservice-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.760683 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-webhook-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.760733 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfgd\" (UniqueName: \"kubernetes.io/projected/62d14162-3eb9-47ac-9c09-b77d858a6318-kube-api-access-8rfgd\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.862178 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-apiservice-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.862237 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-webhook-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.862292 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfgd\" (UniqueName: \"kubernetes.io/projected/62d14162-3eb9-47ac-9c09-b77d858a6318-kube-api-access-8rfgd\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.881982 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-webhook-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.881996 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62d14162-3eb9-47ac-9c09-b77d858a6318-apiservice-cert\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.900014 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfgd\" (UniqueName: \"kubernetes.io/projected/62d14162-3eb9-47ac-9c09-b77d858a6318-kube-api-access-8rfgd\") pod \"elastic-operator-7997b94565-8v5lj\" (UID: \"62d14162-3eb9-47ac-9c09-b77d858a6318\") " pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:20 crc kubenswrapper[4899]: I0123 09:06:20.920269 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7997b94565-8v5lj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.585764 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj"] Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.586842 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.588562 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.588629 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.592770 4899 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-6m7tk" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.601957 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj"] Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.719522 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e0e731fa-506e-48e5-bde2-94098c55a323-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.719597 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6krw\" (UniqueName: \"kubernetes.io/projected/e0e731fa-506e-48e5-bde2-94098c55a323-kube-api-access-m6krw\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.824126 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6krw\" (UniqueName: \"kubernetes.io/projected/e0e731fa-506e-48e5-bde2-94098c55a323-kube-api-access-m6krw\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.824312 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e0e731fa-506e-48e5-bde2-94098c55a323-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.825032 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e0e731fa-506e-48e5-bde2-94098c55a323-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.850147 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6krw\" (UniqueName: \"kubernetes.io/projected/e0e731fa-506e-48e5-bde2-94098c55a323-kube-api-access-m6krw\") pod \"cert-manager-operator-controller-manager-5446d6888b-snjjj\" (UID: \"e0e731fa-506e-48e5-bde2-94098c55a323\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:26 crc kubenswrapper[4899]: I0123 09:06:26.901526 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" Jan 23 09:06:29 crc kubenswrapper[4899]: E0123 09:06:29.456627 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a" Jan 23 09:06:29 crc kubenswrapper[4899]: E0123 09:06:29.457007 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator --watch-referenced-objects-in-all-namespaces=true --disable-unmanaged-prometheus-configuration=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:9a2097bc5b2e02bc1703f64c452ce8fe4bc6775b732db930ff4770b76ae4653a,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lb57g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-68bc856cb9-5dpg7_openshift-operators(a9bf5caa-85e0-4b70-8145-d6c25fa15999): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 09:06:29 crc kubenswrapper[4899]: E0123 09:06:29.458429 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" podUID="a9bf5caa-85e0-4b70-8145-d6c25fa15999" Jan 23 09:06:30 crc kubenswrapper[4899]: E0123 09:06:30.448591 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a\\\"\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" podUID="a9bf5caa-85e0-4b70-8145-d6c25fa15999" Jan 23 09:06:31 crc kubenswrapper[4899]: E0123 09:06:31.335455 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8" Jan 23 09:06:31 crc kubenswrapper[4899]: E0123 09:06:31.335647 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nw78x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5bf474d74f-vgmkw_openshift-operators(28fbc9c0-2fdd-46d3-bc64-bcec54804e17): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 09:06:31 crc kubenswrapper[4899]: E0123 09:06:31.336836 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" podUID="28fbc9c0-2fdd-46d3-bc64-bcec54804e17" Jan 23 09:06:31 crc kubenswrapper[4899]: E0123 09:06:31.454246 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8\\\"\"" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" podUID="28fbc9c0-2fdd-46d3-bc64-bcec54804e17" Jan 23 09:06:33 crc kubenswrapper[4899]: E0123 09:06:33.726018 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 23 09:06:33 crc kubenswrapper[4899]: E0123 09:06:33.726400 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2xtg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-44llv_openshift-marketplace(58968c93-e09f-4d90-acf8-4bb499beb6e5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 09:06:33 crc kubenswrapper[4899]: E0123 09:06:33.727567 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-44llv" podUID="58968c93-e09f-4d90-acf8-4bb499beb6e5" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.081068 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7997b94565-8v5lj"] Jan 23 09:06:34 crc kubenswrapper[4899]: W0123 09:06:34.089283 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62d14162_3eb9_47ac_9c09_b77d858a6318.slice/crio-c02e5440a6b3440a312a7778204af2dfbb98ad9dab0a8aa7e34415642e052a00 WatchSource:0}: Error finding container c02e5440a6b3440a312a7778204af2dfbb98ad9dab0a8aa7e34415642e052a00: Status 404 returned error can't find the container with id c02e5440a6b3440a312a7778204af2dfbb98ad9dab0a8aa7e34415642e052a00 Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.251126 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj"] Jan 23 09:06:34 crc kubenswrapper[4899]: W0123 09:06:34.251141 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0e731fa_506e_48e5_bde2_94098c55a323.slice/crio-9b477195685ddd7865e1d8322d19123da52d930d87665df19f1dfbb470d98933 WatchSource:0}: Error finding container 9b477195685ddd7865e1d8322d19123da52d930d87665df19f1dfbb470d98933: Status 404 returned error can't find the container with id 9b477195685ddd7865e1d8322d19123da52d930d87665df19f1dfbb470d98933 Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.477687 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" event={"ID":"6a152f52-cbf3-43d4-b9fb-0b312c9c0737","Type":"ContainerStarted","Data":"2fa33320f2b5097384fd761d909be4a15cf81e98fb8dd44227aae1d123949f5c"} Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.478310 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.479235 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" event={"ID":"e0e731fa-506e-48e5-bde2-94098c55a323","Type":"ContainerStarted","Data":"9b477195685ddd7865e1d8322d19123da52d930d87665df19f1dfbb470d98933"} Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.483264 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" event={"ID":"ab9f7afb-bc13-43ce-b218-bdb18a103b51","Type":"ContainerStarted","Data":"1e3c1b0075d6207f86556c1b1632917ce9d2c567f56284a748799a35c56e13cc"} Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.485436 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" event={"ID":"e90996a1-43d9-44c4-b23e-e8b110764326","Type":"ContainerStarted","Data":"10d27a39a8f0cf5a76e46b5e54fdb61ca4c0c8f9304bc0d6ad7f8f15cf994959"} Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.486774 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7997b94565-8v5lj" event={"ID":"62d14162-3eb9-47ac-9c09-b77d858a6318","Type":"ContainerStarted","Data":"c02e5440a6b3440a312a7778204af2dfbb98ad9dab0a8aa7e34415642e052a00"} Jan 23 09:06:34 crc kubenswrapper[4899]: E0123 09:06:34.488316 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-44llv" podUID="58968c93-e09f-4d90-acf8-4bb499beb6e5" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.498923 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.506009 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-g9rtj" podStartSLOduration=1.732567561 podStartE2EDuration="23.50598369s" podCreationTimestamp="2026-01-23 09:06:11 +0000 UTC" firstStartedPulling="2026-01-23 09:06:12.022417494 +0000 UTC m=+858.889387881" lastFinishedPulling="2026-01-23 09:06:33.795833623 +0000 UTC m=+880.662804010" observedRunningTime="2026-01-23 09:06:34.49669499 +0000 UTC m=+881.363665397" watchObservedRunningTime="2026-01-23 09:06:34.50598369 +0000 UTC m=+881.372954077" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.554389 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd" podStartSLOduration=1.766466362 podStartE2EDuration="23.554370181s" podCreationTimestamp="2026-01-23 09:06:11 +0000 UTC" firstStartedPulling="2026-01-23 09:06:11.997573676 +0000 UTC m=+858.864544063" lastFinishedPulling="2026-01-23 09:06:33.785477495 +0000 UTC m=+880.652447882" observedRunningTime="2026-01-23 09:06:34.550289191 +0000 UTC m=+881.417259588" watchObservedRunningTime="2026-01-23 09:06:34.554370181 +0000 UTC m=+881.421340568" Jan 23 09:06:34 crc kubenswrapper[4899]: I0123 09:06:34.576759 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw" podStartSLOduration=1.552540109 podStartE2EDuration="23.576732042s" podCreationTimestamp="2026-01-23 09:06:11 +0000 UTC" firstStartedPulling="2026-01-23 09:06:11.781997879 +0000 UTC m=+858.648968266" lastFinishedPulling="2026-01-23 09:06:33.806189812 +0000 UTC m=+880.673160199" observedRunningTime="2026-01-23 09:06:34.574971555 +0000 UTC m=+881.441941942" watchObservedRunningTime="2026-01-23 09:06:34.576732042 +0000 UTC m=+881.443702449" Jan 23 09:06:36 crc kubenswrapper[4899]: I0123 09:06:36.502304 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7997b94565-8v5lj" event={"ID":"62d14162-3eb9-47ac-9c09-b77d858a6318","Type":"ContainerStarted","Data":"ad9f985ed39b5684c629307bf53ecaf9239686439afe080334e7827a919097c4"} Jan 23 09:06:36 crc kubenswrapper[4899]: I0123 09:06:36.536671 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-7997b94565-8v5lj" podStartSLOduration=14.368596224 podStartE2EDuration="16.536642544s" podCreationTimestamp="2026-01-23 09:06:20 +0000 UTC" firstStartedPulling="2026-01-23 09:06:34.093525118 +0000 UTC m=+880.960495505" lastFinishedPulling="2026-01-23 09:06:36.261571438 +0000 UTC m=+883.128541825" observedRunningTime="2026-01-23 09:06:36.521633291 +0000 UTC m=+883.388603698" watchObservedRunningTime="2026-01-23 09:06:36.536642544 +0000 UTC m=+883.403612931" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.485240 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.492166 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.496816 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.497046 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.497305 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-hl669" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.497461 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.498422 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.498584 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.498740 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.499368 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.504725 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.506324 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618202 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618265 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618291 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618315 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618364 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618385 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618403 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618429 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618445 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618497 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618514 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618532 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618550 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618565 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.618589 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/bff1497a-cc11-4f67-9f66-3fba05dceef1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720041 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720093 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720112 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720144 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720163 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720184 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720200 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720231 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720254 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/bff1497a-cc11-4f67-9f66-3fba05dceef1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720276 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720295 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720313 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720330 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720353 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.720377 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.721785 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.722446 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.722686 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.724408 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.724427 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.724616 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.724654 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.724720 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.730892 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.730948 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.730993 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.731184 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.731756 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/bff1497a-cc11-4f67-9f66-3fba05dceef1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.732056 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.737149 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/bff1497a-cc11-4f67-9f66-3fba05dceef1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"bff1497a-cc11-4f67-9f66-3fba05dceef1\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:37 crc kubenswrapper[4899]: I0123 09:06:37.814887 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:06:38 crc kubenswrapper[4899]: I0123 09:06:38.181318 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 23 09:06:38 crc kubenswrapper[4899]: I0123 09:06:38.515919 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"bff1497a-cc11-4f67-9f66-3fba05dceef1","Type":"ContainerStarted","Data":"ffe7d05625256066b34800af31c7e20d0fd31c00cbb05eedfeeb3b790c35170a"} Jan 23 09:06:43 crc kubenswrapper[4899]: I0123 09:06:43.559075 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" event={"ID":"a9bf5caa-85e0-4b70-8145-d6c25fa15999","Type":"ContainerStarted","Data":"cc127e796282c207657e4008a255aca7d6aa442f27ef6ce0fbf9522ecca241f5"} Jan 23 09:06:43 crc kubenswrapper[4899]: I0123 09:06:43.562029 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" event={"ID":"e0e731fa-506e-48e5-bde2-94098c55a323","Type":"ContainerStarted","Data":"783436dae698ba010b4a35ceeba65f14f848ac0dd6caef9f5dbe1ea8074de9ff"} Jan 23 09:06:43 crc kubenswrapper[4899]: I0123 09:06:43.630586 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5dpg7" podStartSLOduration=2.649162559 podStartE2EDuration="33.630568561s" podCreationTimestamp="2026-01-23 09:06:10 +0000 UTC" firstStartedPulling="2026-01-23 09:06:11.492921977 +0000 UTC m=+858.359892364" lastFinishedPulling="2026-01-23 09:06:42.474327979 +0000 UTC m=+889.341298366" observedRunningTime="2026-01-23 09:06:43.599593278 +0000 UTC m=+890.466563665" watchObservedRunningTime="2026-01-23 09:06:43.630568561 +0000 UTC m=+890.497538948" Jan 23 09:06:43 crc kubenswrapper[4899]: I0123 09:06:43.632634 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-snjjj" podStartSLOduration=9.321697944 podStartE2EDuration="17.632627176s" podCreationTimestamp="2026-01-23 09:06:26 +0000 UTC" firstStartedPulling="2026-01-23 09:06:34.25388231 +0000 UTC m=+881.120852697" lastFinishedPulling="2026-01-23 09:06:42.564811542 +0000 UTC m=+889.431781929" observedRunningTime="2026-01-23 09:06:43.629126812 +0000 UTC m=+890.496097229" watchObservedRunningTime="2026-01-23 09:06:43.632627176 +0000 UTC m=+890.499597563" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.629975 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" event={"ID":"28fbc9c0-2fdd-46d3-bc64-bcec54804e17","Type":"ContainerStarted","Data":"6dd626e413d3f397299dccce4721c516533e27c33c9a18c755c91efff2eb57c8"} Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.631372 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.660553 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" podStartSLOduration=2.028702173 podStartE2EDuration="37.660534376s" podCreationTimestamp="2026-01-23 09:06:11 +0000 UTC" firstStartedPulling="2026-01-23 09:06:12.118334883 +0000 UTC m=+858.985305270" lastFinishedPulling="2026-01-23 09:06:47.750167096 +0000 UTC m=+894.617137473" observedRunningTime="2026-01-23 09:06:48.65954993 +0000 UTC m=+895.526520317" watchObservedRunningTime="2026-01-23 09:06:48.660534376 +0000 UTC m=+895.527504763" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.965021 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt"] Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.965831 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.968620 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.968805 4899 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-dqbrh" Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.973143 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt"] Jan 23 09:06:48 crc kubenswrapper[4899]: I0123 09:06:48.975590 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.051322 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.051363 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v78hq\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-kube-api-access-v78hq\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.376338 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.376388 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v78hq\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-kube-api-access-v78hq\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.426982 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.442905 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v78hq\" (UniqueName: \"kubernetes.io/projected/db949b26-d7d6-45e7-abd0-70953e7a713d-kube-api-access-v78hq\") pod \"cert-manager-cainjector-855d9ccff4-6zpnt\" (UID: \"db949b26-d7d6-45e7-abd0-70953e7a713d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.579996 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" Jan 23 09:06:49 crc kubenswrapper[4899]: I0123 09:06:49.934055 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt"] Jan 23 09:06:49 crc kubenswrapper[4899]: W0123 09:06:49.944654 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb949b26_d7d6_45e7_abd0_70953e7a713d.slice/crio-ad379a66a443192f759a7a6e2c842e4e917dd99107714c83e0e7239783d1797c WatchSource:0}: Error finding container ad379a66a443192f759a7a6e2c842e4e917dd99107714c83e0e7239783d1797c: Status 404 returned error can't find the container with id ad379a66a443192f759a7a6e2c842e4e917dd99107714c83e0e7239783d1797c Jan 23 09:06:50 crc kubenswrapper[4899]: I0123 09:06:50.645794 4899 generic.go:334] "Generic (PLEG): container finished" podID="58968c93-e09f-4d90-acf8-4bb499beb6e5" containerID="4fce78a27912f4ea7e882d5961c544e4e3d03382c9cad7c3800f0a976261f1ca" exitCode=0 Jan 23 09:06:50 crc kubenswrapper[4899]: I0123 09:06:50.645877 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44llv" event={"ID":"58968c93-e09f-4d90-acf8-4bb499beb6e5","Type":"ContainerDied","Data":"4fce78a27912f4ea7e882d5961c544e4e3d03382c9cad7c3800f0a976261f1ca"} Jan 23 09:06:50 crc kubenswrapper[4899]: I0123 09:06:50.650853 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" event={"ID":"db949b26-d7d6-45e7-abd0-70953e7a713d","Type":"ContainerStarted","Data":"ad379a66a443192f759a7a6e2c842e4e917dd99107714c83e0e7239783d1797c"} Jan 23 09:06:51 crc kubenswrapper[4899]: I0123 09:06:51.816634 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44llv" event={"ID":"58968c93-e09f-4d90-acf8-4bb499beb6e5","Type":"ContainerStarted","Data":"d35583d45d03943336e336ff16aaf96aa0b1a55f625ab57b049b6e0ed124bdc3"} Jan 23 09:06:51 crc kubenswrapper[4899]: I0123 09:06:51.851465 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-44llv" podStartSLOduration=8.827074972 podStartE2EDuration="47.85142541s" podCreationTimestamp="2026-01-23 09:06:04 +0000 UTC" firstStartedPulling="2026-01-23 09:06:12.24239716 +0000 UTC m=+859.109367547" lastFinishedPulling="2026-01-23 09:06:51.266747598 +0000 UTC m=+898.133717985" observedRunningTime="2026-01-23 09:06:51.833872038 +0000 UTC m=+898.700842425" watchObservedRunningTime="2026-01-23 09:06:51.85142541 +0000 UTC m=+898.718395797" Jan 23 09:06:55 crc kubenswrapper[4899]: I0123 09:06:55.083010 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:55 crc kubenswrapper[4899]: I0123 09:06:55.083792 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:55 crc kubenswrapper[4899]: I0123 09:06:55.267865 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.759090 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-xbh7b"] Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.760269 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.767461 4899 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-r7drj" Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.779140 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-xbh7b"] Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.915462 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvb9w\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-kube-api-access-wvb9w\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:56 crc kubenswrapper[4899]: I0123 09:06:56.915535 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:57 crc kubenswrapper[4899]: I0123 09:06:57.017117 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvb9w\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-kube-api-access-wvb9w\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:57 crc kubenswrapper[4899]: I0123 09:06:57.017183 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:57 crc kubenswrapper[4899]: I0123 09:06:57.047114 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:57 crc kubenswrapper[4899]: I0123 09:06:57.047266 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvb9w\" (UniqueName: \"kubernetes.io/projected/e0d0ee40-1af4-479a-84f9-87322410c93f-kube-api-access-wvb9w\") pod \"cert-manager-webhook-f4fb5df64-xbh7b\" (UID: \"e0d0ee40-1af4-479a-84f9-87322410c93f\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:06:57 crc kubenswrapper[4899]: I0123 09:06:57.167671 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.311993 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.313323 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.324625 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.406444 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.406518 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kkdt\" (UniqueName: \"kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.406540 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.507557 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.507648 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kkdt\" (UniqueName: \"kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.507675 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.508430 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.508536 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.530226 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kkdt\" (UniqueName: \"kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt\") pod \"community-operators-ldxq7\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.629350 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:01 crc kubenswrapper[4899]: I0123 09:07:01.795481 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-vgmkw" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.216582 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-44llv" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.314704 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zmxfr"] Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.315618 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.318804 4899 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-t4fw2" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.337261 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44llv"] Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.345174 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zmxfr"] Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.487332 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnf2h\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-kube-api-access-hnf2h\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.487724 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-bound-sa-token\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.575203 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.575468 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vtljc" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" containerID="cri-o://ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" gracePeriod=2 Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.588812 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnf2h\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-kube-api-access-hnf2h\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.589470 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-bound-sa-token\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.762586 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnf2h\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-kube-api-access-hnf2h\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.767110 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/067ef4a9-ba7d-4d41-8b04-c3180e74c354-bound-sa-token\") pod \"cert-manager-86cb77c54b-zmxfr\" (UID: \"067ef4a9-ba7d-4d41-8b04-c3180e74c354\") " pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:05 crc kubenswrapper[4899]: I0123 09:07:05.934768 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-zmxfr" Jan 23 09:07:06 crc kubenswrapper[4899]: E0123 09:07:06.683194 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:06 crc kubenswrapper[4899]: E0123 09:07:06.684522 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:06 crc kubenswrapper[4899]: E0123 09:07:06.684818 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:06 crc kubenswrapper[4899]: E0123 09:07:06.684891 4899 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-vtljc" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" Jan 23 09:07:08 crc kubenswrapper[4899]: I0123 09:07:08.037708 4899 generic.go:334] "Generic (PLEG): container finished" podID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" exitCode=0 Jan 23 09:07:08 crc kubenswrapper[4899]: I0123 09:07:08.037791 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerDied","Data":"ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5"} Jan 23 09:07:16 crc kubenswrapper[4899]: E0123 09:07:16.682878 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:16 crc kubenswrapper[4899]: E0123 09:07:16.683747 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:16 crc kubenswrapper[4899]: E0123 09:07:16.684116 4899 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" cmd=["grpc_health_probe","-addr=:50051"] Jan 23 09:07:16 crc kubenswrapper[4899]: E0123 09:07:16.684153 4899 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-vtljc" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.183555 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.184563 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-cainjector,Image:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,Command:[/app/cmd/cainjector/cainjector],Args:[--leader-election-namespace=kube-system --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bound-sa-token,ReadOnly:true,MountPath:/var/run/secrets/openshift/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v78hq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-cainjector-855d9ccff4-6zpnt_cert-manager(db949b26-d7d6-45e7-abd0-70953e7a713d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.185653 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-cainjector\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" podUID="db949b26-d7d6-45e7-abd0-70953e7a713d" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.533325 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.582015 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.582311 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(bff1497a-cc11-4f67-9f66-3fba05dceef1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:07:22 crc kubenswrapper[4899]: E0123 09:07:22.583528 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.664637 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content\") pod \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.664706 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn2fs\" (UniqueName: \"kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs\") pod \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.664754 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities\") pod \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\" (UID: \"a03a9c2b-6cd4-4ff6-9085-121a461bf730\") " Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.665743 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities" (OuterVolumeSpecName: "utilities") pod "a03a9c2b-6cd4-4ff6-9085-121a461bf730" (UID: "a03a9c2b-6cd4-4ff6-9085-121a461bf730"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.671303 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs" (OuterVolumeSpecName: "kube-api-access-mn2fs") pod "a03a9c2b-6cd4-4ff6-9085-121a461bf730" (UID: "a03a9c2b-6cd4-4ff6-9085-121a461bf730"). InnerVolumeSpecName "kube-api-access-mn2fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.685869 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.731429 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a03a9c2b-6cd4-4ff6-9085-121a461bf730" (UID: "a03a9c2b-6cd4-4ff6-9085-121a461bf730"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.766416 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn2fs\" (UniqueName: \"kubernetes.io/projected/a03a9c2b-6cd4-4ff6-9085-121a461bf730-kube-api-access-mn2fs\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.766465 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.766481 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a03a9c2b-6cd4-4ff6-9085-121a461bf730-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.857636 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zmxfr"] Jan 23 09:07:22 crc kubenswrapper[4899]: W0123 09:07:22.863166 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod067ef4a9_ba7d_4d41_8b04_c3180e74c354.slice/crio-d079731b8126aac7a23403c95083969dc5ebab5921eba95e8d2fd398b6056d6e WatchSource:0}: Error finding container d079731b8126aac7a23403c95083969dc5ebab5921eba95e8d2fd398b6056d6e: Status 404 returned error can't find the container with id d079731b8126aac7a23403c95083969dc5ebab5921eba95e8d2fd398b6056d6e Jan 23 09:07:22 crc kubenswrapper[4899]: I0123 09:07:22.958637 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-xbh7b"] Jan 23 09:07:22 crc kubenswrapper[4899]: W0123 09:07:22.961282 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0d0ee40_1af4_479a_84f9_87322410c93f.slice/crio-5a516fe8af9e2827089ffd541ce6e549867eebffac1248850c7b98f34a794582 WatchSource:0}: Error finding container 5a516fe8af9e2827089ffd541ce6e549867eebffac1248850c7b98f34a794582: Status 404 returned error can't find the container with id 5a516fe8af9e2827089ffd541ce6e549867eebffac1248850c7b98f34a794582 Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.197685 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" event={"ID":"e0d0ee40-1af4-479a-84f9-87322410c93f","Type":"ContainerStarted","Data":"5a516fe8af9e2827089ffd541ce6e549867eebffac1248850c7b98f34a794582"} Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.199976 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vtljc" event={"ID":"a03a9c2b-6cd4-4ff6-9085-121a461bf730","Type":"ContainerDied","Data":"34646bcc3c4d68f1fc010658d60f6cc7494c2226fe1a37d19096623807f58d55"} Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.200009 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vtljc" Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.200034 4899 scope.go:117] "RemoveContainer" containerID="ed13d03b9920dadaadfc3ad9ea8fd575eaffc1557e64e4cd70db07a3123b75d5" Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.201539 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerID="b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403" exitCode=0 Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.201612 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerDied","Data":"b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403"} Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.201640 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerStarted","Data":"a20c52cbcb10c6d9634c88ee8209a79dee2518c61a3f2228181af70e8d288ad2"} Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.203166 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-zmxfr" event={"ID":"067ef4a9-ba7d-4d41-8b04-c3180e74c354","Type":"ContainerStarted","Data":"d079731b8126aac7a23403c95083969dc5ebab5921eba95e8d2fd398b6056d6e"} Jan 23 09:07:23 crc kubenswrapper[4899]: E0123 09:07:23.203669 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-cainjector\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df\\\"\"" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" podUID="db949b26-d7d6-45e7-abd0-70953e7a713d" Jan 23 09:07:23 crc kubenswrapper[4899]: E0123 09:07:23.204144 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.220412 4899 scope.go:117] "RemoveContainer" containerID="896bbdf82fe34a8e0916cde513a6cef6bc2e9a375f2e8f902831ec6c50ad7519" Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.243580 4899 scope.go:117] "RemoveContainer" containerID="acd89e1851a2d5566894f2879d3905d4bcf847232224ba544fa54cc4d44a0f7a" Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.279521 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.285915 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vtljc"] Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.479424 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.510168 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 23 09:07:23 crc kubenswrapper[4899]: I0123 09:07:23.522107 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" path="/var/lib/kubelet/pods/a03a9c2b-6cd4-4ff6-9085-121a461bf730/volumes" Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.210716 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerID="83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280" exitCode=0 Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.210786 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerDied","Data":"83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280"} Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.212394 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-zmxfr" event={"ID":"067ef4a9-ba7d-4d41-8b04-c3180e74c354","Type":"ContainerStarted","Data":"9a649eb8386e5bca39c10958400dbfa26e50f7feab2165458c8104a3dc34f286"} Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.215094 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" event={"ID":"e0d0ee40-1af4-479a-84f9-87322410c93f","Type":"ContainerStarted","Data":"7cc9a376b47292b815ddf16cebf2175a112e5022d7b1e85dbea2454828406144"} Jan 23 09:07:24 crc kubenswrapper[4899]: E0123 09:07:24.216241 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.242523 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-zmxfr" podStartSLOduration=18.551800994 podStartE2EDuration="19.242505747s" podCreationTimestamp="2026-01-23 09:07:05 +0000 UTC" firstStartedPulling="2026-01-23 09:07:22.86511548 +0000 UTC m=+929.732085867" lastFinishedPulling="2026-01-23 09:07:23.555820233 +0000 UTC m=+930.422790620" observedRunningTime="2026-01-23 09:07:24.241685384 +0000 UTC m=+931.108655781" watchObservedRunningTime="2026-01-23 09:07:24.242505747 +0000 UTC m=+931.109476134" Jan 23 09:07:24 crc kubenswrapper[4899]: I0123 09:07:24.255082 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" podStartSLOduration=27.49327807 podStartE2EDuration="28.255059663s" podCreationTimestamp="2026-01-23 09:06:56 +0000 UTC" firstStartedPulling="2026-01-23 09:07:22.965806576 +0000 UTC m=+929.832776973" lastFinishedPulling="2026-01-23 09:07:23.727588179 +0000 UTC m=+930.594558566" observedRunningTime="2026-01-23 09:07:24.253733416 +0000 UTC m=+931.120703823" watchObservedRunningTime="2026-01-23 09:07:24.255059663 +0000 UTC m=+931.122030050" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.228503 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerStarted","Data":"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef"} Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.229172 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:07:25 crc kubenswrapper[4899]: E0123 09:07:25.232268 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.249271 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ldxq7" podStartSLOduration=22.538724622 podStartE2EDuration="24.249242145s" podCreationTimestamp="2026-01-23 09:07:01 +0000 UTC" firstStartedPulling="2026-01-23 09:07:23.203861569 +0000 UTC m=+930.070831956" lastFinishedPulling="2026-01-23 09:07:24.914379082 +0000 UTC m=+931.781349479" observedRunningTime="2026-01-23 09:07:25.245164092 +0000 UTC m=+932.112134489" watchObservedRunningTime="2026-01-23 09:07:25.249242145 +0000 UTC m=+932.116212552" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.408625 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Jan 23 09:07:25 crc kubenswrapper[4899]: E0123 09:07:25.408908 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.408922 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" Jan 23 09:07:25 crc kubenswrapper[4899]: E0123 09:07:25.408938 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="extract-utilities" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.408946 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="extract-utilities" Jan 23 09:07:25 crc kubenswrapper[4899]: E0123 09:07:25.408957 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="extract-content" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.408963 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="extract-content" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.409076 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03a9c2b-6cd4-4ff6-9085-121a461bf730" containerName="registry-server" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.409846 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.412809 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.413026 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-sw4dm" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.413091 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.413276 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.413463 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.439597 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509434 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509479 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509499 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509529 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509546 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-sw4dm-pull\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509632 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509667 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509819 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9khtp\" (UniqueName: \"kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509840 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509856 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509873 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509898 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.509916 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-sw4dm-push\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.610968 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9khtp\" (UniqueName: \"kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611023 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611045 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611066 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611101 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611121 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-sw4dm-push\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611140 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611160 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611180 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611237 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611256 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-sw4dm-pull\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611274 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.611299 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612171 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612298 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612326 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612381 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612585 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612675 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612816 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.612959 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.613440 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.616798 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.616942 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-sw4dm-push\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.617963 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-sw4dm-pull\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.634933 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9khtp\" (UniqueName: \"kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp\") pod \"service-telemetry-framework-index-1-build\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.729777 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:07:25 crc kubenswrapper[4899]: I0123 09:07:25.978612 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Jan 23 09:07:26 crc kubenswrapper[4899]: I0123 09:07:26.235580 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerStarted","Data":"fb778963ff38f55343f8e20d4a49bd8a59929279dfba6390380f9c617dd1f7c1"} Jan 23 09:07:31 crc kubenswrapper[4899]: I0123 09:07:31.630449 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:31 crc kubenswrapper[4899]: I0123 09:07:31.630499 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:31 crc kubenswrapper[4899]: I0123 09:07:31.665956 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:31 crc kubenswrapper[4899]: I0123 09:07:31.801027 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:07:31 crc kubenswrapper[4899]: I0123 09:07:31.801080 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:07:32 crc kubenswrapper[4899]: I0123 09:07:32.171346 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-xbh7b" Jan 23 09:07:32 crc kubenswrapper[4899]: I0123 09:07:32.304113 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerStarted","Data":"4a38e4d1df3df31a20a295d7ca4d6b73c2f6782cea33fe72f1fffc048502f9aa"} Jan 23 09:07:32 crc kubenswrapper[4899]: I0123 09:07:32.339179 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:32 crc kubenswrapper[4899]: I0123 09:07:32.516539 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:33 crc kubenswrapper[4899]: I0123 09:07:33.316941 4899 generic.go:334] "Generic (PLEG): container finished" podID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerID="4a38e4d1df3df31a20a295d7ca4d6b73c2f6782cea33fe72f1fffc048502f9aa" exitCode=0 Jan 23 09:07:33 crc kubenswrapper[4899]: I0123 09:07:33.317017 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerDied","Data":"4a38e4d1df3df31a20a295d7ca4d6b73c2f6782cea33fe72f1fffc048502f9aa"} Jan 23 09:07:34 crc kubenswrapper[4899]: I0123 09:07:34.325145 4899 generic.go:334] "Generic (PLEG): container finished" podID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerID="45ab0876da93a6812aa39cc6f31b219ef6b736f2f7e6590e1324aa12c8294309" exitCode=0 Jan 23 09:07:34 crc kubenswrapper[4899]: I0123 09:07:34.325240 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerDied","Data":"45ab0876da93a6812aa39cc6f31b219ef6b736f2f7e6590e1324aa12c8294309"} Jan 23 09:07:34 crc kubenswrapper[4899]: I0123 09:07:34.325540 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ldxq7" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="registry-server" containerID="cri-o://7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef" gracePeriod=2 Jan 23 09:07:34 crc kubenswrapper[4899]: I0123 09:07:34.359050 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_9e6a9dcb-da33-4b2c-8c9f-a402042a5720/manage-dockerfile/0.log" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.312775 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.341264 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerStarted","Data":"dd70e8bd895a5f22ba1da9e2684082bad1d6403e8e045dbd04791cf22cb8d4c5"} Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.349468 4899 generic.go:334] "Generic (PLEG): container finished" podID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerID="7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef" exitCode=0 Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.349523 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerDied","Data":"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef"} Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.349564 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldxq7" event={"ID":"f4d66347-3915-45a8-9d22-5acece6a6bd4","Type":"ContainerDied","Data":"a20c52cbcb10c6d9634c88ee8209a79dee2518c61a3f2228181af70e8d288ad2"} Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.349586 4899 scope.go:117] "RemoveContainer" containerID="7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.349737 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldxq7" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.379518 4899 scope.go:117] "RemoveContainer" containerID="83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.387001 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-framework-index-1-build" podStartSLOduration=4.428024752 podStartE2EDuration="10.386977213s" podCreationTimestamp="2026-01-23 09:07:25 +0000 UTC" firstStartedPulling="2026-01-23 09:07:25.987218452 +0000 UTC m=+932.854188839" lastFinishedPulling="2026-01-23 09:07:31.946170913 +0000 UTC m=+938.813141300" observedRunningTime="2026-01-23 09:07:35.375690362 +0000 UTC m=+942.242660749" watchObservedRunningTime="2026-01-23 09:07:35.386977213 +0000 UTC m=+942.253947600" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.387154 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kkdt\" (UniqueName: \"kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt\") pod \"f4d66347-3915-45a8-9d22-5acece6a6bd4\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.387269 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities\") pod \"f4d66347-3915-45a8-9d22-5acece6a6bd4\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.387422 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content\") pod \"f4d66347-3915-45a8-9d22-5acece6a6bd4\" (UID: \"f4d66347-3915-45a8-9d22-5acece6a6bd4\") " Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.389072 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities" (OuterVolumeSpecName: "utilities") pod "f4d66347-3915-45a8-9d22-5acece6a6bd4" (UID: "f4d66347-3915-45a8-9d22-5acece6a6bd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.393171 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt" (OuterVolumeSpecName: "kube-api-access-8kkdt") pod "f4d66347-3915-45a8-9d22-5acece6a6bd4" (UID: "f4d66347-3915-45a8-9d22-5acece6a6bd4"). InnerVolumeSpecName "kube-api-access-8kkdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.449599 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4d66347-3915-45a8-9d22-5acece6a6bd4" (UID: "f4d66347-3915-45a8-9d22-5acece6a6bd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.488838 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.488874 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kkdt\" (UniqueName: \"kubernetes.io/projected/f4d66347-3915-45a8-9d22-5acece6a6bd4-kube-api-access-8kkdt\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.488884 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d66347-3915-45a8-9d22-5acece6a6bd4-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.675026 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:35 crc kubenswrapper[4899]: I0123 09:07:35.680663 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ldxq7"] Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.572599 4899 scope.go:117] "RemoveContainer" containerID="b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.607634 4899 scope.go:117] "RemoveContainer" containerID="7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef" Jan 23 09:07:36 crc kubenswrapper[4899]: E0123 09:07:36.608096 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef\": container with ID starting with 7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef not found: ID does not exist" containerID="7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.608147 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef"} err="failed to get container status \"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef\": rpc error: code = NotFound desc = could not find container \"7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef\": container with ID starting with 7c6e086febd32bc0e21988ffa06555767049ab085f6cda85424d1045e08ab1ef not found: ID does not exist" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.608187 4899 scope.go:117] "RemoveContainer" containerID="83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280" Jan 23 09:07:36 crc kubenswrapper[4899]: E0123 09:07:36.608574 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280\": container with ID starting with 83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280 not found: ID does not exist" containerID="83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.608607 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280"} err="failed to get container status \"83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280\": rpc error: code = NotFound desc = could not find container \"83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280\": container with ID starting with 83eaa2de5f9bbe7d8a6257bf39be623e6480327437a1fe6db4d8a6c73016e280 not found: ID does not exist" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.608623 4899 scope.go:117] "RemoveContainer" containerID="b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403" Jan 23 09:07:36 crc kubenswrapper[4899]: E0123 09:07:36.608978 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403\": container with ID starting with b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403 not found: ID does not exist" containerID="b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403" Jan 23 09:07:36 crc kubenswrapper[4899]: I0123 09:07:36.609006 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403"} err="failed to get container status \"b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403\": rpc error: code = NotFound desc = could not find container \"b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403\": container with ID starting with b22d2397efef41c5cec3b5036a36cef08e1bdcd10b7acddb76f3cce67ec69403 not found: ID does not exist" Jan 23 09:07:37 crc kubenswrapper[4899]: I0123 09:07:37.522765 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" path="/var/lib/kubelet/pods/f4d66347-3915-45a8-9d22-5acece6a6bd4/volumes" Jan 23 09:07:41 crc kubenswrapper[4899]: I0123 09:07:41.394108 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" event={"ID":"db949b26-d7d6-45e7-abd0-70953e7a713d","Type":"ContainerStarted","Data":"4648118c7a34e526a10b0a88a1f3d6b2ccec79f6347fb2941e93238c3fe71f96"} Jan 23 09:07:41 crc kubenswrapper[4899]: I0123 09:07:41.411863 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6zpnt" podStartSLOduration=-9223371983.44293 podStartE2EDuration="53.411846532s" podCreationTimestamp="2026-01-23 09:06:48 +0000 UTC" firstStartedPulling="2026-01-23 09:06:49.947161864 +0000 UTC m=+896.814132251" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 09:07:41.408723285 +0000 UTC m=+948.275693692" watchObservedRunningTime="2026-01-23 09:07:41.411846532 +0000 UTC m=+948.278816919" Jan 23 09:07:42 crc kubenswrapper[4899]: I0123 09:07:42.400132 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"bff1497a-cc11-4f67-9f66-3fba05dceef1","Type":"ContainerStarted","Data":"39690364696e6506fa189ca61da771346dcec1ee20d87bf05a5ca6428b82138e"} Jan 23 09:07:43 crc kubenswrapper[4899]: I0123 09:07:43.458171 4899 generic.go:334] "Generic (PLEG): container finished" podID="bff1497a-cc11-4f67-9f66-3fba05dceef1" containerID="39690364696e6506fa189ca61da771346dcec1ee20d87bf05a5ca6428b82138e" exitCode=0 Jan 23 09:07:43 crc kubenswrapper[4899]: I0123 09:07:43.458252 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"bff1497a-cc11-4f67-9f66-3fba05dceef1","Type":"ContainerDied","Data":"39690364696e6506fa189ca61da771346dcec1ee20d87bf05a5ca6428b82138e"} Jan 23 09:07:44 crc kubenswrapper[4899]: I0123 09:07:44.476552 4899 generic.go:334] "Generic (PLEG): container finished" podID="bff1497a-cc11-4f67-9f66-3fba05dceef1" containerID="c7438a9ba82036fa8f47e65a58f2812e212c371c6bd42bd038963ff6de1dee87" exitCode=0 Jan 23 09:07:44 crc kubenswrapper[4899]: I0123 09:07:44.476610 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"bff1497a-cc11-4f67-9f66-3fba05dceef1","Type":"ContainerDied","Data":"c7438a9ba82036fa8f47e65a58f2812e212c371c6bd42bd038963ff6de1dee87"} Jan 23 09:07:45 crc kubenswrapper[4899]: I0123 09:07:45.483959 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"bff1497a-cc11-4f67-9f66-3fba05dceef1","Type":"ContainerStarted","Data":"1808fb48c8c69977e926d6997926fb9237f81554bba293c5001e8681c1aa39ac"} Jan 23 09:07:45 crc kubenswrapper[4899]: I0123 09:07:45.484362 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:07:45 crc kubenswrapper[4899]: I0123 09:07:45.522368 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=5.645891398 podStartE2EDuration="1m8.522347707s" podCreationTimestamp="2026-01-23 09:06:37 +0000 UTC" firstStartedPulling="2026-01-23 09:06:38.179423108 +0000 UTC m=+885.046393495" lastFinishedPulling="2026-01-23 09:07:41.055879417 +0000 UTC m=+947.922849804" observedRunningTime="2026-01-23 09:07:45.514981454 +0000 UTC m=+952.381951841" watchObservedRunningTime="2026-01-23 09:07:45.522347707 +0000 UTC m=+952.389318094" Jan 23 09:07:57 crc kubenswrapper[4899]: I0123 09:07:57.946017 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" containerName="elasticsearch" probeResult="failure" output=< Jan 23 09:07:57 crc kubenswrapper[4899]: {"timestamp": "2026-01-23T09:07:57+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 23 09:07:57 crc kubenswrapper[4899]: > Jan 23 09:08:01 crc kubenswrapper[4899]: I0123 09:08:01.685860 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:08:01 crc kubenswrapper[4899]: I0123 09:08:01.686184 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:08:03 crc kubenswrapper[4899]: I0123 09:08:03.100379 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" containerName="elasticsearch" probeResult="failure" output=< Jan 23 09:08:03 crc kubenswrapper[4899]: {"timestamp": "2026-01-23T09:08:03+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 23 09:08:03 crc kubenswrapper[4899]: > Jan 23 09:08:07 crc kubenswrapper[4899]: I0123 09:08:07.963051 4899 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="bff1497a-cc11-4f67-9f66-3fba05dceef1" containerName="elasticsearch" probeResult="failure" output=< Jan 23 09:08:07 crc kubenswrapper[4899]: {"timestamp": "2026-01-23T09:08:07+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 23 09:08:07 crc kubenswrapper[4899]: > Jan 23 09:08:13 crc kubenswrapper[4899]: I0123 09:08:13.315738 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 23 09:08:31 crc kubenswrapper[4899]: I0123 09:08:31.686697 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:08:31 crc kubenswrapper[4899]: I0123 09:08:31.688143 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:08:31 crc kubenswrapper[4899]: I0123 09:08:31.688350 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:08:31 crc kubenswrapper[4899]: I0123 09:08:31.690175 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:08:31 crc kubenswrapper[4899]: I0123 09:08:31.690416 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41" gracePeriod=600 Jan 23 09:08:32 crc kubenswrapper[4899]: I0123 09:08:32.062002 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41" exitCode=0 Jan 23 09:08:32 crc kubenswrapper[4899]: I0123 09:08:32.062040 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41"} Jan 23 09:08:32 crc kubenswrapper[4899]: I0123 09:08:32.062068 4899 scope.go:117] "RemoveContainer" containerID="490154d22b2b60697973b0cc83a78a7a90570db2f4f799a162c1320a4739b6c4" Jan 23 09:08:33 crc kubenswrapper[4899]: I0123 09:08:33.070694 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6"} Jan 23 09:08:34 crc kubenswrapper[4899]: I0123 09:08:34.080807 4899 generic.go:334] "Generic (PLEG): container finished" podID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerID="dd70e8bd895a5f22ba1da9e2684082bad1d6403e8e045dbd04791cf22cb8d4c5" exitCode=0 Jan 23 09:08:34 crc kubenswrapper[4899]: I0123 09:08:34.080880 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerDied","Data":"dd70e8bd895a5f22ba1da9e2684082bad1d6403e8e045dbd04791cf22cb8d4c5"} Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.351760 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458068 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458166 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458222 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458293 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9khtp\" (UniqueName: \"kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458344 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-sw4dm-pull\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458388 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-sw4dm-push\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458421 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458462 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458511 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458563 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458581 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458611 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.458643 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run\") pod \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\" (UID: \"9e6a9dcb-da33-4b2c-8c9f-a402042a5720\") " Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.473883 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.474341 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.475051 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.475463 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.478295 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.479316 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.483410 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.485772 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp" (OuterVolumeSpecName: "kube-api-access-9khtp") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "kube-api-access-9khtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.492436 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull" (OuterVolumeSpecName: "builder-dockercfg-sw4dm-pull") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "builder-dockercfg-sw4dm-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.492531 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push" (OuterVolumeSpecName: "builder-dockercfg-sw4dm-push") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "builder-dockercfg-sw4dm-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.493771 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561888 4899 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561928 4899 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561939 4899 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561951 4899 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561963 4899 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561973 4899 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561982 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9khtp\" (UniqueName: \"kubernetes.io/projected/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-kube-api-access-9khtp\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.561993 4899 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-sw4dm-pull\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-pull\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.562003 4899 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-sw4dm-push\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-builder-dockercfg-sw4dm-push\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.562013 4899 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.562024 4899 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.774366 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:08:35 crc kubenswrapper[4899]: I0123 09:08:35.867852 4899 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.094611 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"9e6a9dcb-da33-4b2c-8c9f-a402042a5720","Type":"ContainerDied","Data":"fb778963ff38f55343f8e20d4a49bd8a59929279dfba6390380f9c617dd1f7c1"} Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.094656 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb778963ff38f55343f8e20d4a49bd8a59929279dfba6390380f9c617dd1f7c1" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.094945 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.555649 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9e6a9dcb-da33-4b2c-8c9f-a402042a5720" (UID: "9e6a9dcb-da33-4b2c-8c9f-a402042a5720"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.577586 4899 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9e6a9dcb-da33-4b2c-8c9f-a402042a5720-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733339 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733753 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="git-clone" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733764 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="git-clone" Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733776 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="manage-dockerfile" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733782 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="manage-dockerfile" Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733791 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="registry-server" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733797 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="registry-server" Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733810 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="extract-utilities" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733817 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="extract-utilities" Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733833 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="docker-build" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733839 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="docker-build" Jan 23 09:08:36 crc kubenswrapper[4899]: E0123 09:08:36.733848 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="extract-content" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733855 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="extract-content" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733945 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6a9dcb-da33-4b2c-8c9f-a402042a5720" containerName="docker-build" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.733956 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d66347-3915-45a8-9d22-5acece6a6bd4" containerName="registry-server" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.734407 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.739760 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-5v7vn" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.753500 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.880444 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7hm9\" (UniqueName: \"kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9\") pod \"infrawatch-operators-psz2h\" (UID: \"93588e62-5de1-4ebf-976d-48a467748bc2\") " pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:36 crc kubenswrapper[4899]: I0123 09:08:36.981765 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7hm9\" (UniqueName: \"kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9\") pod \"infrawatch-operators-psz2h\" (UID: \"93588e62-5de1-4ebf-976d-48a467748bc2\") " pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:37 crc kubenswrapper[4899]: I0123 09:08:37.000610 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7hm9\" (UniqueName: \"kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9\") pod \"infrawatch-operators-psz2h\" (UID: \"93588e62-5de1-4ebf-976d-48a467748bc2\") " pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:37 crc kubenswrapper[4899]: I0123 09:08:37.056772 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:37 crc kubenswrapper[4899]: I0123 09:08:37.261271 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:38 crc kubenswrapper[4899]: I0123 09:08:38.110878 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-psz2h" event={"ID":"93588e62-5de1-4ebf-976d-48a467748bc2","Type":"ContainerStarted","Data":"87f25ed49755223313ab6cfe7c653214e6c2eeef363395fd2982e3b578b0f8ba"} Jan 23 09:08:41 crc kubenswrapper[4899]: I0123 09:08:41.107943 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:41 crc kubenswrapper[4899]: I0123 09:08:41.916996 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-g9fts"] Jan 23 09:08:41 crc kubenswrapper[4899]: I0123 09:08:41.922812 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:41 crc kubenswrapper[4899]: I0123 09:08:41.933754 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g9fts"] Jan 23 09:08:42 crc kubenswrapper[4899]: I0123 09:08:42.050776 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56chd\" (UniqueName: \"kubernetes.io/projected/c6833739-119b-4334-9d63-1a38956e8b2a-kube-api-access-56chd\") pod \"infrawatch-operators-g9fts\" (UID: \"c6833739-119b-4334-9d63-1a38956e8b2a\") " pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:42 crc kubenswrapper[4899]: I0123 09:08:42.152592 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56chd\" (UniqueName: \"kubernetes.io/projected/c6833739-119b-4334-9d63-1a38956e8b2a-kube-api-access-56chd\") pod \"infrawatch-operators-g9fts\" (UID: \"c6833739-119b-4334-9d63-1a38956e8b2a\") " pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:42 crc kubenswrapper[4899]: I0123 09:08:42.173211 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56chd\" (UniqueName: \"kubernetes.io/projected/c6833739-119b-4334-9d63-1a38956e8b2a-kube-api-access-56chd\") pod \"infrawatch-operators-g9fts\" (UID: \"c6833739-119b-4334-9d63-1a38956e8b2a\") " pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:42 crc kubenswrapper[4899]: I0123 09:08:42.250915 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:47 crc kubenswrapper[4899]: I0123 09:08:47.698834 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g9fts"] Jan 23 09:08:51 crc kubenswrapper[4899]: I0123 09:08:51.184825 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g9fts" event={"ID":"c6833739-119b-4334-9d63-1a38956e8b2a","Type":"ContainerStarted","Data":"cb94b34afe1c77f0cfb5376ce968fbdbaed4d69a9277adcaf8d5c13ff4a59caa"} Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.191703 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-psz2h" event={"ID":"93588e62-5de1-4ebf-976d-48a467748bc2","Type":"ContainerStarted","Data":"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031"} Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.191797 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-psz2h" podUID="93588e62-5de1-4ebf-976d-48a467748bc2" containerName="registry-server" containerID="cri-o://783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031" gracePeriod=2 Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.194618 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g9fts" event={"ID":"c6833739-119b-4334-9d63-1a38956e8b2a","Type":"ContainerStarted","Data":"9c0adcbc94b139afd77b21e9498b7f280042a195861a101d21fd0cfe9ef9ca82"} Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.216485 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-psz2h" podStartSLOduration=2.215481638 podStartE2EDuration="16.216466994s" podCreationTimestamp="2026-01-23 09:08:36 +0000 UTC" firstStartedPulling="2026-01-23 09:08:37.270827231 +0000 UTC m=+1004.137797618" lastFinishedPulling="2026-01-23 09:08:51.271812587 +0000 UTC m=+1018.138782974" observedRunningTime="2026-01-23 09:08:52.212797852 +0000 UTC m=+1019.079768239" watchObservedRunningTime="2026-01-23 09:08:52.216466994 +0000 UTC m=+1019.083437381" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.229227 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-g9fts" podStartSLOduration=10.35967107 podStartE2EDuration="11.229187304s" podCreationTimestamp="2026-01-23 09:08:41 +0000 UTC" firstStartedPulling="2026-01-23 09:08:50.338667959 +0000 UTC m=+1017.205638356" lastFinishedPulling="2026-01-23 09:08:51.208184203 +0000 UTC m=+1018.075154590" observedRunningTime="2026-01-23 09:08:52.228640529 +0000 UTC m=+1019.095610916" watchObservedRunningTime="2026-01-23 09:08:52.229187304 +0000 UTC m=+1019.096157691" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.251909 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.251957 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.277659 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.600381 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.726510 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7hm9\" (UniqueName: \"kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9\") pod \"93588e62-5de1-4ebf-976d-48a467748bc2\" (UID: \"93588e62-5de1-4ebf-976d-48a467748bc2\") " Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.734096 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9" (OuterVolumeSpecName: "kube-api-access-g7hm9") pod "93588e62-5de1-4ebf-976d-48a467748bc2" (UID: "93588e62-5de1-4ebf-976d-48a467748bc2"). InnerVolumeSpecName "kube-api-access-g7hm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:08:52 crc kubenswrapper[4899]: I0123 09:08:52.828085 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7hm9\" (UniqueName: \"kubernetes.io/projected/93588e62-5de1-4ebf-976d-48a467748bc2-kube-api-access-g7hm9\") on node \"crc\" DevicePath \"\"" Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.202492 4899 generic.go:334] "Generic (PLEG): container finished" podID="93588e62-5de1-4ebf-976d-48a467748bc2" containerID="783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031" exitCode=0 Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.202555 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-psz2h" Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.202587 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-psz2h" event={"ID":"93588e62-5de1-4ebf-976d-48a467748bc2","Type":"ContainerDied","Data":"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031"} Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.202633 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-psz2h" event={"ID":"93588e62-5de1-4ebf-976d-48a467748bc2","Type":"ContainerDied","Data":"87f25ed49755223313ab6cfe7c653214e6c2eeef363395fd2982e3b578b0f8ba"} Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.202651 4899 scope.go:117] "RemoveContainer" containerID="783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031" Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.229258 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.230297 4899 scope.go:117] "RemoveContainer" containerID="783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031" Jan 23 09:08:53 crc kubenswrapper[4899]: E0123 09:08:53.230667 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031\": container with ID starting with 783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031 not found: ID does not exist" containerID="783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031" Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.230705 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031"} err="failed to get container status \"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031\": rpc error: code = NotFound desc = could not find container \"783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031\": container with ID starting with 783e7b03f57bc79b5afe20cd73e069072290384ca504fa3408c9809038d3e031 not found: ID does not exist" Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.232920 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-psz2h"] Jan 23 09:08:53 crc kubenswrapper[4899]: I0123 09:08:53.524739 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93588e62-5de1-4ebf-976d-48a467748bc2" path="/var/lib/kubelet/pods/93588e62-5de1-4ebf-976d-48a467748bc2/volumes" Jan 23 09:09:02 crc kubenswrapper[4899]: I0123 09:09:02.280450 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-g9fts" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.743251 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7"] Jan 23 09:09:10 crc kubenswrapper[4899]: E0123 09:09:10.743984 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93588e62-5de1-4ebf-976d-48a467748bc2" containerName="registry-server" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.743999 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="93588e62-5de1-4ebf-976d-48a467748bc2" containerName="registry-server" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.744127 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="93588e62-5de1-4ebf-976d-48a467748bc2" containerName="registry-server" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.745123 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.748774 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.750531 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7"] Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.765909 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.765958 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw5zz\" (UniqueName: \"kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.766077 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.867445 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw5zz\" (UniqueName: \"kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.867581 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.867632 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.868124 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.868239 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:10 crc kubenswrapper[4899]: I0123 09:09:10.886794 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw5zz\" (UniqueName: \"kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.068947 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.506439 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7"] Jan 23 09:09:11 crc kubenswrapper[4899]: W0123 09:09:11.510770 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee87fff6_abaf_4b8b_ace1_d2440eb834ed.slice/crio-4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27 WatchSource:0}: Error finding container 4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27: Status 404 returned error can't find the container with id 4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27 Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.622800 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88"] Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.624234 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.640292 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88"] Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.678025 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjpjh\" (UniqueName: \"kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.678104 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.678159 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.779341 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.779445 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjpjh\" (UniqueName: \"kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.779498 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.779888 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.779970 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.799040 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjpjh\" (UniqueName: \"kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:11 crc kubenswrapper[4899]: I0123 09:09:11.943317 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.199647 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88"] Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.320348 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" event={"ID":"d098fdbc-0bb3-459f-8156-3f17bb75e503","Type":"ContainerStarted","Data":"718a1b1cf3ea228cd62d5f7f9729abf6b53710e2a1c2833faf9f6eae21409329"} Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.321860 4899 generic.go:334] "Generic (PLEG): container finished" podID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerID="a6a647cd7c629d63908162d2e64fd76699d1041fb3cf0e86ae1192c1196696bb" exitCode=0 Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.321900 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" event={"ID":"ee87fff6-abaf-4b8b-ace1-d2440eb834ed","Type":"ContainerDied","Data":"a6a647cd7c629d63908162d2e64fd76699d1041fb3cf0e86ae1192c1196696bb"} Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.321924 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" event={"ID":"ee87fff6-abaf-4b8b-ace1-d2440eb834ed","Type":"ContainerStarted","Data":"4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27"} Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.602103 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct"] Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.603329 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.611941 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct"] Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.688965 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdkdt\" (UniqueName: \"kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.689107 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.689141 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.790115 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdkdt\" (UniqueName: \"kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.790200 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.790253 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.790774 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.790818 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.810303 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdkdt\" (UniqueName: \"kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:12 crc kubenswrapper[4899]: I0123 09:09:12.964808 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:13 crc kubenswrapper[4899]: I0123 09:09:13.183130 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct"] Jan 23 09:09:13 crc kubenswrapper[4899]: I0123 09:09:13.329949 4899 generic.go:334] "Generic (PLEG): container finished" podID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerID="3596b268aecb6cd2b5568cf6987511b5b60f017f6c70bbd8bad551261de1d73d" exitCode=0 Jan 23 09:09:13 crc kubenswrapper[4899]: I0123 09:09:13.330030 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" event={"ID":"d098fdbc-0bb3-459f-8156-3f17bb75e503","Type":"ContainerDied","Data":"3596b268aecb6cd2b5568cf6987511b5b60f017f6c70bbd8bad551261de1d73d"} Jan 23 09:09:13 crc kubenswrapper[4899]: I0123 09:09:13.334125 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" event={"ID":"de2da763-0d3d-435e-a80d-94eada1deff7","Type":"ContainerStarted","Data":"78370f0a2328483ca9e5637f07990962c4919391f42a559e51bbd46d05ef2a2a"} Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.340278 4899 generic.go:334] "Generic (PLEG): container finished" podID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerID="ae67a5bf7d07c548ea0e4c6836058d5b63d797e8b191939665b2a09ffd1fa777" exitCode=0 Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.340591 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" event={"ID":"ee87fff6-abaf-4b8b-ace1-d2440eb834ed","Type":"ContainerDied","Data":"ae67a5bf7d07c548ea0e4c6836058d5b63d797e8b191939665b2a09ffd1fa777"} Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.341997 4899 generic.go:334] "Generic (PLEG): container finished" podID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerID="2acdc5dca84f14c7b0398a915d8ea8b3d8fdfbb680fa94a1e430376751034b2e" exitCode=0 Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.342060 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" event={"ID":"d098fdbc-0bb3-459f-8156-3f17bb75e503","Type":"ContainerDied","Data":"2acdc5dca84f14c7b0398a915d8ea8b3d8fdfbb680fa94a1e430376751034b2e"} Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.345008 4899 generic.go:334] "Generic (PLEG): container finished" podID="de2da763-0d3d-435e-a80d-94eada1deff7" containerID="27812c0c4a70c867b1626d8bfd8cbf407a04b2d67fed4752951a4931582484b6" exitCode=0 Jan 23 09:09:14 crc kubenswrapper[4899]: I0123 09:09:14.345058 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" event={"ID":"de2da763-0d3d-435e-a80d-94eada1deff7","Type":"ContainerDied","Data":"27812c0c4a70c867b1626d8bfd8cbf407a04b2d67fed4752951a4931582484b6"} Jan 23 09:09:15 crc kubenswrapper[4899]: I0123 09:09:15.354123 4899 generic.go:334] "Generic (PLEG): container finished" podID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerID="8f14d8376733586c4ad4610c958abcacdc9c0c8d43ed5fe3e5de64bfb240426c" exitCode=0 Jan 23 09:09:15 crc kubenswrapper[4899]: I0123 09:09:15.354500 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" event={"ID":"ee87fff6-abaf-4b8b-ace1-d2440eb834ed","Type":"ContainerDied","Data":"8f14d8376733586c4ad4610c958abcacdc9c0c8d43ed5fe3e5de64bfb240426c"} Jan 23 09:09:15 crc kubenswrapper[4899]: I0123 09:09:15.357295 4899 generic.go:334] "Generic (PLEG): container finished" podID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerID="7fe90a2aca7ecb8ed3abb24eb32b4ed5a6fa152642c99ebb437a6cef976b754b" exitCode=0 Jan 23 09:09:15 crc kubenswrapper[4899]: I0123 09:09:15.357343 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" event={"ID":"d098fdbc-0bb3-459f-8156-3f17bb75e503","Type":"ContainerDied","Data":"7fe90a2aca7ecb8ed3abb24eb32b4ed5a6fa152642c99ebb437a6cef976b754b"} Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.284396 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.359451 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.371874 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" event={"ID":"d098fdbc-0bb3-459f-8156-3f17bb75e503","Type":"ContainerDied","Data":"718a1b1cf3ea228cd62d5f7f9729abf6b53710e2a1c2833faf9f6eae21409329"} Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.371906 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7ebv2b88" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.371916 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="718a1b1cf3ea228cd62d5f7f9729abf6b53710e2a1c2833faf9f6eae21409329" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.376979 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" event={"ID":"ee87fff6-abaf-4b8b-ace1-d2440eb834ed","Type":"ContainerDied","Data":"4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27"} Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.377020 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d2e4ca1f6c72c29478e1d5257d2a44b98ca50ef92669fcbc60f3f6ac4ddbf27" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.377072 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.447851 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw5zz\" (UniqueName: \"kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz\") pod \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.447926 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle\") pod \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.448028 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util\") pod \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\" (UID: \"ee87fff6-abaf-4b8b-ace1-d2440eb834ed\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.448761 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle" (OuterVolumeSpecName: "bundle") pod "ee87fff6-abaf-4b8b-ace1-d2440eb834ed" (UID: "ee87fff6-abaf-4b8b-ace1-d2440eb834ed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.452803 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz" (OuterVolumeSpecName: "kube-api-access-cw5zz") pod "ee87fff6-abaf-4b8b-ace1-d2440eb834ed" (UID: "ee87fff6-abaf-4b8b-ace1-d2440eb834ed"). InnerVolumeSpecName "kube-api-access-cw5zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.549756 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util\") pod \"d098fdbc-0bb3-459f-8156-3f17bb75e503\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.549828 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle\") pod \"d098fdbc-0bb3-459f-8156-3f17bb75e503\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.549879 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjpjh\" (UniqueName: \"kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh\") pod \"d098fdbc-0bb3-459f-8156-3f17bb75e503\" (UID: \"d098fdbc-0bb3-459f-8156-3f17bb75e503\") " Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.550125 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw5zz\" (UniqueName: \"kubernetes.io/projected/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-kube-api-access-cw5zz\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.550157 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.550601 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle" (OuterVolumeSpecName: "bundle") pod "d098fdbc-0bb3-459f-8156-3f17bb75e503" (UID: "d098fdbc-0bb3-459f-8156-3f17bb75e503"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.552752 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh" (OuterVolumeSpecName: "kube-api-access-sjpjh") pod "d098fdbc-0bb3-459f-8156-3f17bb75e503" (UID: "d098fdbc-0bb3-459f-8156-3f17bb75e503"). InnerVolumeSpecName "kube-api-access-sjpjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.564817 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util" (OuterVolumeSpecName: "util") pod "d098fdbc-0bb3-459f-8156-3f17bb75e503" (UID: "d098fdbc-0bb3-459f-8156-3f17bb75e503"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.593960 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util" (OuterVolumeSpecName: "util") pod "ee87fff6-abaf-4b8b-ace1-d2440eb834ed" (UID: "ee87fff6-abaf-4b8b-ace1-d2440eb834ed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.651148 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.651185 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee87fff6-abaf-4b8b-ace1-d2440eb834ed-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.651199 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d098fdbc-0bb3-459f-8156-3f17bb75e503-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:17 crc kubenswrapper[4899]: I0123 09:09:17.651225 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjpjh\" (UniqueName: \"kubernetes.io/projected/d098fdbc-0bb3-459f-8156-3f17bb75e503-kube-api-access-sjpjh\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:18 crc kubenswrapper[4899]: I0123 09:09:18.385743 4899 generic.go:334] "Generic (PLEG): container finished" podID="de2da763-0d3d-435e-a80d-94eada1deff7" containerID="3f77e25b9196b557bfca84f2349e818addff16d51da7a4368b6afda15865b81a" exitCode=0 Jan 23 09:09:18 crc kubenswrapper[4899]: I0123 09:09:18.385826 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" event={"ID":"de2da763-0d3d-435e-a80d-94eada1deff7","Type":"ContainerDied","Data":"3f77e25b9196b557bfca84f2349e818addff16d51da7a4368b6afda15865b81a"} Jan 23 09:09:19 crc kubenswrapper[4899]: I0123 09:09:19.392822 4899 generic.go:334] "Generic (PLEG): container finished" podID="de2da763-0d3d-435e-a80d-94eada1deff7" containerID="98061236c4eb10447568dfba40b3a9debce69446351b18f292c6ea0034d9ed74" exitCode=0 Jan 23 09:09:19 crc kubenswrapper[4899]: I0123 09:09:19.392929 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" event={"ID":"de2da763-0d3d-435e-a80d-94eada1deff7","Type":"ContainerDied","Data":"98061236c4eb10447568dfba40b3a9debce69446351b18f292c6ea0034d9ed74"} Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.656532 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.788480 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util\") pod \"de2da763-0d3d-435e-a80d-94eada1deff7\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.788618 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle\") pod \"de2da763-0d3d-435e-a80d-94eada1deff7\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.788729 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdkdt\" (UniqueName: \"kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt\") pod \"de2da763-0d3d-435e-a80d-94eada1deff7\" (UID: \"de2da763-0d3d-435e-a80d-94eada1deff7\") " Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.789287 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle" (OuterVolumeSpecName: "bundle") pod "de2da763-0d3d-435e-a80d-94eada1deff7" (UID: "de2da763-0d3d-435e-a80d-94eada1deff7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.790525 4899 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-bundle\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.794154 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt" (OuterVolumeSpecName: "kube-api-access-gdkdt") pod "de2da763-0d3d-435e-a80d-94eada1deff7" (UID: "de2da763-0d3d-435e-a80d-94eada1deff7"). InnerVolumeSpecName "kube-api-access-gdkdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.800623 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util" (OuterVolumeSpecName: "util") pod "de2da763-0d3d-435e-a80d-94eada1deff7" (UID: "de2da763-0d3d-435e-a80d-94eada1deff7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.891578 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdkdt\" (UniqueName: \"kubernetes.io/projected/de2da763-0d3d-435e-a80d-94eada1deff7-kube-api-access-gdkdt\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:20 crc kubenswrapper[4899]: I0123 09:09:20.891876 4899 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de2da763-0d3d-435e-a80d-94eada1deff7-util\") on node \"crc\" DevicePath \"\"" Jan 23 09:09:21 crc kubenswrapper[4899]: I0123 09:09:21.406184 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" event={"ID":"de2da763-0d3d-435e-a80d-94eada1deff7","Type":"ContainerDied","Data":"78370f0a2328483ca9e5637f07990962c4919391f42a559e51bbd46d05ef2a2a"} Jan 23 09:09:21 crc kubenswrapper[4899]: I0123 09:09:21.406463 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78370f0a2328483ca9e5637f07990962c4919391f42a559e51bbd46d05ef2a2a" Jan 23 09:09:21 crc kubenswrapper[4899]: I0123 09:09:21.406314 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bbzxct" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131337 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv"] Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131848 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131865 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131875 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131883 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131896 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131903 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131917 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131924 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="pull" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131931 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131937 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131947 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131953 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131962 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131968 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="util" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.131982 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.131989 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: E0123 09:09:25.132002 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.132009 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.132121 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee87fff6-abaf-4b8b-ace1-d2440eb834ed" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.132138 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2da763-0d3d-435e-a80d-94eada1deff7" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.132146 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="d098fdbc-0bb3-459f-8156-3f17bb75e503" containerName="extract" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.132631 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.139333 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-vh96m" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.141122 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv"] Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.248013 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9d6x\" (UniqueName: \"kubernetes.io/projected/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-kube-api-access-l9d6x\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.248097 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-runner\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.349764 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-runner\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.350171 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9d6x\" (UniqueName: \"kubernetes.io/projected/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-kube-api-access-l9d6x\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.350272 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-runner\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.369352 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9d6x\" (UniqueName: \"kubernetes.io/projected/1ca0b316-d4d5-4c10-aff5-24bd8fde1d49-kube-api-access-l9d6x\") pod \"service-telemetry-operator-55b89ddfb9-jgdbv\" (UID: \"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49\") " pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.490768 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" Jan 23 09:09:25 crc kubenswrapper[4899]: I0123 09:09:25.710792 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv"] Jan 23 09:09:26 crc kubenswrapper[4899]: I0123 09:09:26.432726 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" event={"ID":"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49","Type":"ContainerStarted","Data":"723e23c6f8f9c43d6fe1900a7147c44b863b00e9c4a141f794fd18a31afe72d7"} Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.692903 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vf8qs"] Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.694096 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.696855 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-q7rfw" Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.703622 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vf8qs"] Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.828605 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7884w\" (UniqueName: \"kubernetes.io/projected/549c2897-bf7e-400e-8b11-e6f13982a4f7-kube-api-access-7884w\") pod \"interconnect-operator-5bb49f789d-vf8qs\" (UID: \"549c2897-bf7e-400e-8b11-e6f13982a4f7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.930163 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7884w\" (UniqueName: \"kubernetes.io/projected/549c2897-bf7e-400e-8b11-e6f13982a4f7-kube-api-access-7884w\") pod \"interconnect-operator-5bb49f789d-vf8qs\" (UID: \"549c2897-bf7e-400e-8b11-e6f13982a4f7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" Jan 23 09:09:27 crc kubenswrapper[4899]: I0123 09:09:27.947705 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7884w\" (UniqueName: \"kubernetes.io/projected/549c2897-bf7e-400e-8b11-e6f13982a4f7-kube-api-access-7884w\") pod \"interconnect-operator-5bb49f789d-vf8qs\" (UID: \"549c2897-bf7e-400e-8b11-e6f13982a4f7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.015497 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.333097 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-bbbc889bc-dgksc"] Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.334304 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.341083 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-dw68w" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.341560 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bbbc889bc-dgksc"] Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.435829 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/566e53d3-281e-4452-a072-d387390d427c-runner\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.435893 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6tzm\" (UniqueName: \"kubernetes.io/projected/566e53d3-281e-4452-a072-d387390d427c-kube-api-access-m6tzm\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.473271 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vf8qs"] Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.532051 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" event={"ID":"549c2897-bf7e-400e-8b11-e6f13982a4f7","Type":"ContainerStarted","Data":"d2d5343371e84c11f5dc59a98c5b0bd7ee8b4bd3119028576fdcf16d2d0636df"} Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.536965 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/566e53d3-281e-4452-a072-d387390d427c-runner\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.537040 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6tzm\" (UniqueName: \"kubernetes.io/projected/566e53d3-281e-4452-a072-d387390d427c-kube-api-access-m6tzm\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.537639 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/566e53d3-281e-4452-a072-d387390d427c-runner\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.562471 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6tzm\" (UniqueName: \"kubernetes.io/projected/566e53d3-281e-4452-a072-d387390d427c-kube-api-access-m6tzm\") pod \"smart-gateway-operator-bbbc889bc-dgksc\" (UID: \"566e53d3-281e-4452-a072-d387390d427c\") " pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.654026 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" Jan 23 09:09:28 crc kubenswrapper[4899]: I0123 09:09:28.860113 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bbbc889bc-dgksc"] Jan 23 09:09:28 crc kubenswrapper[4899]: W0123 09:09:28.871707 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566e53d3_281e_4452_a072_d387390d427c.slice/crio-6ff5af92e20c3ad3265eee31ad9684be0534a50835c9f3ecfbd77ff7dd7d2464 WatchSource:0}: Error finding container 6ff5af92e20c3ad3265eee31ad9684be0534a50835c9f3ecfbd77ff7dd7d2464: Status 404 returned error can't find the container with id 6ff5af92e20c3ad3265eee31ad9684be0534a50835c9f3ecfbd77ff7dd7d2464 Jan 23 09:09:29 crc kubenswrapper[4899]: I0123 09:09:29.541083 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" event={"ID":"566e53d3-281e-4452-a072-d387390d427c","Type":"ContainerStarted","Data":"6ff5af92e20c3ad3265eee31ad9684be0534a50835c9f3ecfbd77ff7dd7d2464"} Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.776623 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/service-telemetry-operator:latest" Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.778008 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/service-telemetry-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:service-telemetry-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_WEBHOOK_SNMP_IMAGE,Value:quay.io/infrawatch/prometheus-webhook-snmp:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_IMAGE,Value:quay.io/prometheus/prometheus:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER_IMAGE,Value:quay.io/prometheus/alertmanager:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:service-telemetry-operator.v1.5.1768085182,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9d6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod service-telemetry-operator-55b89ddfb9-jgdbv_service-telemetry(1ca0b316-d4d5-4c10-aff5-24bd8fde1d49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.779221 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" podUID="1ca0b316-d4d5-4c10-aff5-24bd8fde1d49" Jan 23 09:09:58 crc kubenswrapper[4899]: I0123 09:09:58.880222 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" event={"ID":"549c2897-bf7e-400e-8b11-e6f13982a4f7","Type":"ContainerStarted","Data":"3bd4f246855356a99485f2786ee2fa67d4deba1b5e1ea4bc193391486bb20601"} Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.881760 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/service-telemetry-operator:latest\\\"\"" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" podUID="1ca0b316-d4d5-4c10-aff5-24bd8fde1d49" Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.882107 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:latest" Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.882297 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1768085178,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6tzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-bbbc889bc-dgksc_service-telemetry(566e53d3-281e-4452-a072-d387390d427c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:09:58 crc kubenswrapper[4899]: E0123 09:09:58.883444 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" podUID="566e53d3-281e-4452-a072-d387390d427c" Jan 23 09:09:58 crc kubenswrapper[4899]: I0123 09:09:58.919006 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-vf8qs" podStartSLOduration=17.990909799 podStartE2EDuration="31.918988388s" podCreationTimestamp="2026-01-23 09:09:27 +0000 UTC" firstStartedPulling="2026-01-23 09:09:28.487245128 +0000 UTC m=+1055.354215515" lastFinishedPulling="2026-01-23 09:09:42.415323717 +0000 UTC m=+1069.282294104" observedRunningTime="2026-01-23 09:09:58.913886617 +0000 UTC m=+1085.780857004" watchObservedRunningTime="2026-01-23 09:09:58.918988388 +0000 UTC m=+1085.785958775" Jan 23 09:10:00 crc kubenswrapper[4899]: E0123 09:10:00.161324 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:latest\\\"\"" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" podUID="566e53d3-281e-4452-a072-d387390d427c" Jan 23 09:10:12 crc kubenswrapper[4899]: I0123 09:10:12.529194 4899 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 23 09:10:14 crc kubenswrapper[4899]: I0123 09:10:14.671991 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" event={"ID":"566e53d3-281e-4452-a072-d387390d427c","Type":"ContainerStarted","Data":"64f688277cf6b928f66155c2f48467c4d1f47108dfd78b5f165ac2e21d8f49fe"} Jan 23 09:10:14 crc kubenswrapper[4899]: I0123 09:10:14.690355 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-bbbc889bc-dgksc" podStartSLOduration=1.03033738 podStartE2EDuration="46.690332696s" podCreationTimestamp="2026-01-23 09:09:28 +0000 UTC" firstStartedPulling="2026-01-23 09:09:28.876893382 +0000 UTC m=+1055.743863769" lastFinishedPulling="2026-01-23 09:10:14.536888698 +0000 UTC m=+1101.403859085" observedRunningTime="2026-01-23 09:10:14.687833177 +0000 UTC m=+1101.554803564" watchObservedRunningTime="2026-01-23 09:10:14.690332696 +0000 UTC m=+1101.557303083" Jan 23 09:10:15 crc kubenswrapper[4899]: I0123 09:10:15.679768 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" event={"ID":"1ca0b316-d4d5-4c10-aff5-24bd8fde1d49","Type":"ContainerStarted","Data":"d609c5feacf6b3ddc9d92e65e8bd1b8e7c64aaf8a167b73920ff5c3b53dab550"} Jan 23 09:10:15 crc kubenswrapper[4899]: I0123 09:10:15.700737 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-55b89ddfb9-jgdbv" podStartSLOduration=1.728061544 podStartE2EDuration="50.700518884s" podCreationTimestamp="2026-01-23 09:09:25 +0000 UTC" firstStartedPulling="2026-01-23 09:09:25.717172336 +0000 UTC m=+1052.584142723" lastFinishedPulling="2026-01-23 09:10:14.689629676 +0000 UTC m=+1101.556600063" observedRunningTime="2026-01-23 09:10:15.695534647 +0000 UTC m=+1102.562505034" watchObservedRunningTime="2026-01-23 09:10:15.700518884 +0000 UTC m=+1102.567489261" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.137363 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.139894 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.142499 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.142503 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.143707 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.143819 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.143856 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.144136 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-xphnz" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.144477 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.151504 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.327595 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.327656 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.327687 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.327722 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.327840 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.328014 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.328074 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2k6n\" (UniqueName: \"kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.428762 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.429114 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2k6n\" (UniqueName: \"kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.429648 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.429792 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.429897 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.430030 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.430140 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.431242 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.435597 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.436179 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.436545 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.446456 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.447622 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2k6n\" (UniqueName: \"kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.452925 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xqn7w\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.461707 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.662472 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:10:40 crc kubenswrapper[4899]: I0123 09:10:40.881001 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" event={"ID":"16f3e64c-a1c5-411e-93bf-a523b3f32bbd","Type":"ContainerStarted","Data":"33c465cd7ecedc0020d67f48af501dfd2f24a160d541bad0bbd41e265f8fc548"} Jan 23 09:10:45 crc kubenswrapper[4899]: I0123 09:10:45.920293 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" event={"ID":"16f3e64c-a1c5-411e-93bf-a523b3f32bbd","Type":"ContainerStarted","Data":"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1"} Jan 23 09:10:45 crc kubenswrapper[4899]: I0123 09:10:45.940553 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" podStartSLOduration=1.120569919 podStartE2EDuration="5.940536833s" podCreationTimestamp="2026-01-23 09:10:40 +0000 UTC" firstStartedPulling="2026-01-23 09:10:40.676804704 +0000 UTC m=+1127.543775091" lastFinishedPulling="2026-01-23 09:10:45.496771618 +0000 UTC m=+1132.363742005" observedRunningTime="2026-01-23 09:10:45.93862474 +0000 UTC m=+1132.805595127" watchObservedRunningTime="2026-01-23 09:10:45.940536833 +0000 UTC m=+1132.807507220" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.544626 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.546185 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553031 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553129 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-q54xg" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553239 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553292 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553376 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553449 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553528 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553527 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553596 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.553659 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.563484 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.662954 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-tls-assets\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663011 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-web-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663041 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663062 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8pq7\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-kube-api-access-p8pq7\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663105 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663128 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663147 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663167 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88d27067-77f9-40fb-970d-df28779d48a2-config-out\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663192 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663251 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663272 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.663300 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.764767 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765103 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765133 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765158 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765244 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-tls-assets\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765270 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-web-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765293 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8pq7\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-kube-api-access-p8pq7\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765312 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765358 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765387 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765406 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.765425 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88d27067-77f9-40fb-970d-df28779d48a2-config-out\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: E0123 09:10:50.766257 4899 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 23 09:10:50 crc kubenswrapper[4899]: E0123 09:10:50.766338 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls podName:88d27067-77f9-40fb-970d-df28779d48a2 nodeName:}" failed. No retries permitted until 2026-01-23 09:10:51.266302915 +0000 UTC m=+1138.133273302 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "88d27067-77f9-40fb-970d-df28779d48a2") : secret "default-prometheus-proxy-tls" not found Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.767456 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.767534 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.768252 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.768472 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/88d27067-77f9-40fb-970d-df28779d48a2-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.788632 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88d27067-77f9-40fb-970d-df28779d48a2-config-out\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.788688 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-tls-assets\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.788730 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.805686 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-web-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.808170 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8pq7\" (UniqueName: \"kubernetes.io/projected/88d27067-77f9-40fb-970d-df28779d48a2-kube-api-access-p8pq7\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.817825 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-config\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.837612 4899 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.837666 4899 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/544624a1ab41ba3e14b999e502141d60e8800e773d6a5d2ea2be7469fa5ccecf/globalmount\"" pod="service-telemetry/prometheus-default-0" Jan 23 09:10:50 crc kubenswrapper[4899]: I0123 09:10:50.883943 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2aaeced3-2a76-427b-9d78-53aef7d6a9a1\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:51 crc kubenswrapper[4899]: I0123 09:10:51.271617 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:51 crc kubenswrapper[4899]: E0123 09:10:51.271807 4899 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 23 09:10:51 crc kubenswrapper[4899]: E0123 09:10:51.271865 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls podName:88d27067-77f9-40fb-970d-df28779d48a2 nodeName:}" failed. No retries permitted until 2026-01-23 09:10:52.271850382 +0000 UTC m=+1139.138820769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "88d27067-77f9-40fb-970d-df28779d48a2") : secret "default-prometheus-proxy-tls" not found Jan 23 09:10:52 crc kubenswrapper[4899]: I0123 09:10:52.284717 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:52 crc kubenswrapper[4899]: I0123 09:10:52.340357 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/88d27067-77f9-40fb-970d-df28779d48a2-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"88d27067-77f9-40fb-970d-df28779d48a2\") " pod="service-telemetry/prometheus-default-0" Jan 23 09:10:52 crc kubenswrapper[4899]: I0123 09:10:52.366290 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 23 09:10:52 crc kubenswrapper[4899]: I0123 09:10:52.805339 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 23 09:10:52 crc kubenswrapper[4899]: I0123 09:10:52.976255 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerStarted","Data":"2e0db5a9a64654f3b98d03c428f8112c87ae7ed4bb52b8216fc4ed24d3f237ef"} Jan 23 09:10:58 crc kubenswrapper[4899]: I0123 09:10:58.024849 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerStarted","Data":"253a561bcb76e17a30597df01e7f1f2aa951718206bac51775f6b5cd4bcf6a23"} Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.171429 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q"] Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.172549 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.180100 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q"] Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.307175 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ckkx\" (UniqueName: \"kubernetes.io/projected/930899a8-bdeb-4546-a142-370356b282bd-kube-api-access-6ckkx\") pod \"default-snmp-webhook-78bcbbdcff-4n76q\" (UID: \"930899a8-bdeb-4546-a142-370356b282bd\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.408840 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ckkx\" (UniqueName: \"kubernetes.io/projected/930899a8-bdeb-4546-a142-370356b282bd-kube-api-access-6ckkx\") pod \"default-snmp-webhook-78bcbbdcff-4n76q\" (UID: \"930899a8-bdeb-4546-a142-370356b282bd\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.428537 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ckkx\" (UniqueName: \"kubernetes.io/projected/930899a8-bdeb-4546-a142-370356b282bd-kube-api-access-6ckkx\") pod \"default-snmp-webhook-78bcbbdcff-4n76q\" (UID: \"930899a8-bdeb-4546-a142-370356b282bd\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.489764 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.685855 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.686408 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:11:01 crc kubenswrapper[4899]: I0123 09:11:01.908010 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q"] Jan 23 09:11:02 crc kubenswrapper[4899]: I0123 09:11:02.049547 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" event={"ID":"930899a8-bdeb-4546-a142-370356b282bd","Type":"ContainerStarted","Data":"6e2e7c59bb7da26022ad71a5f0fbe3292c308059dbb72d3004f3e127e67708c7"} Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.306366 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.312315 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.314918 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.315179 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-947sq" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.320417 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.320479 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.320619 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.364774 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.365988 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.465501 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.465881 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-volume\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.465924 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-out\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.465963 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.465990 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-web-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.466014 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.466035 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.466065 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zdrn\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-kube-api-access-7zdrn\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.466085 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-tls-assets\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567175 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567276 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-volume\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567335 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-out\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567373 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567411 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-web-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567439 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567466 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567500 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zdrn\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-kube-api-access-7zdrn\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.567519 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-tls-assets\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: E0123 09:11:04.567531 4899 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:04 crc kubenswrapper[4899]: E0123 09:11:04.567598 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls podName:ca3cd04b-86ad-4eaf-a48e-1e985eb010b6 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:05.06757746 +0000 UTC m=+1151.934547847 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "ca3cd04b-86ad-4eaf-a48e-1e985eb010b6") : secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.574302 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-web-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.574309 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.575290 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-out\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.577802 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-tls-assets\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.580467 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.581026 4899 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.581092 4899 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/09c23ecc834d5f3aa8b7743d0dc5c6998ce87a7e9bd1537c9b6f046178aaacb7/globalmount\"" pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.581382 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-config-volume\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.591570 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zdrn\" (UniqueName: \"kubernetes.io/projected/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-kube-api-access-7zdrn\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:04 crc kubenswrapper[4899]: I0123 09:11:04.609730 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef847323-f1b3-4c3a-8449-76fa7f82e531\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:05 crc kubenswrapper[4899]: I0123 09:11:05.073919 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:05 crc kubenswrapper[4899]: E0123 09:11:05.074146 4899 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:05 crc kubenswrapper[4899]: E0123 09:11:05.074190 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls podName:ca3cd04b-86ad-4eaf-a48e-1e985eb010b6 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:06.074176375 +0000 UTC m=+1152.941146762 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "ca3cd04b-86ad-4eaf-a48e-1e985eb010b6") : secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:06 crc kubenswrapper[4899]: I0123 09:11:06.114702 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:06 crc kubenswrapper[4899]: E0123 09:11:06.114890 4899 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:06 crc kubenswrapper[4899]: E0123 09:11:06.115174 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls podName:ca3cd04b-86ad-4eaf-a48e-1e985eb010b6 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:08.115154103 +0000 UTC m=+1154.982124490 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "ca3cd04b-86ad-4eaf-a48e-1e985eb010b6") : secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:08 crc kubenswrapper[4899]: I0123 09:11:08.143284 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:08 crc kubenswrapper[4899]: E0123 09:11:08.143442 4899 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:08 crc kubenswrapper[4899]: E0123 09:11:08.143490 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls podName:ca3cd04b-86ad-4eaf-a48e-1e985eb010b6 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:12.143476231 +0000 UTC m=+1159.010446618 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "ca3cd04b-86ad-4eaf-a48e-1e985eb010b6") : secret "default-alertmanager-proxy-tls" not found Jan 23 09:11:09 crc kubenswrapper[4899]: I0123 09:11:09.102447 4899 generic.go:334] "Generic (PLEG): container finished" podID="88d27067-77f9-40fb-970d-df28779d48a2" containerID="253a561bcb76e17a30597df01e7f1f2aa951718206bac51775f6b5cd4bcf6a23" exitCode=0 Jan 23 09:11:09 crc kubenswrapper[4899]: I0123 09:11:09.102542 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerDied","Data":"253a561bcb76e17a30597df01e7f1f2aa951718206bac51775f6b5cd4bcf6a23"} Jan 23 09:11:12 crc kubenswrapper[4899]: I0123 09:11:12.151490 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:12 crc kubenswrapper[4899]: I0123 09:11:12.157937 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca3cd04b-86ad-4eaf-a48e-1e985eb010b6-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6\") " pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:12 crc kubenswrapper[4899]: I0123 09:11:12.228675 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 23 09:11:12 crc kubenswrapper[4899]: I0123 09:11:12.998364 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 23 09:11:13 crc kubenswrapper[4899]: W0123 09:11:13.109036 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca3cd04b_86ad_4eaf_a48e_1e985eb010b6.slice/crio-79ab5156bfd0d95bf4d5fe6e68c6e9f04808cf521c58bdca9169fe73f88b52f4 WatchSource:0}: Error finding container 79ab5156bfd0d95bf4d5fe6e68c6e9f04808cf521c58bdca9169fe73f88b52f4: Status 404 returned error can't find the container with id 79ab5156bfd0d95bf4d5fe6e68c6e9f04808cf521c58bdca9169fe73f88b52f4 Jan 23 09:11:13 crc kubenswrapper[4899]: I0123 09:11:13.152018 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerStarted","Data":"79ab5156bfd0d95bf4d5fe6e68c6e9f04808cf521c58bdca9169fe73f88b52f4"} Jan 23 09:11:13 crc kubenswrapper[4899]: I0123 09:11:13.154320 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" event={"ID":"930899a8-bdeb-4546-a142-370356b282bd","Type":"ContainerStarted","Data":"96e683165076a063bfb6347ddf1a4ac4491c45afabcc9677271e503fc1108f22"} Jan 23 09:11:13 crc kubenswrapper[4899]: I0123 09:11:13.175173 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-4n76q" podStartSLOduration=1.362131524 podStartE2EDuration="12.175151266s" podCreationTimestamp="2026-01-23 09:11:01 +0000 UTC" firstStartedPulling="2026-01-23 09:11:01.916441985 +0000 UTC m=+1148.783412362" lastFinishedPulling="2026-01-23 09:11:12.729461707 +0000 UTC m=+1159.596432104" observedRunningTime="2026-01-23 09:11:13.165744427 +0000 UTC m=+1160.032714824" watchObservedRunningTime="2026-01-23 09:11:13.175151266 +0000 UTC m=+1160.042121663" Jan 23 09:11:16 crc kubenswrapper[4899]: I0123 09:11:16.175870 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerStarted","Data":"0e8ddc812be64895b77686e00073a30a49a24ef71aca16da97e5dff781bd2d0f"} Jan 23 09:11:23 crc kubenswrapper[4899]: I0123 09:11:23.267517 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerStarted","Data":"aa91afd0399608eab946e689918496362ebf72fbe549915f9c2bbacb1a1bfe13"} Jan 23 09:11:24 crc kubenswrapper[4899]: I0123 09:11:24.276554 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerStarted","Data":"be117a8257dc98bc413e02ffd4d9337225c4ac085bc3a53c1778dbdb8086e3e3"} Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.593098 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8"] Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.594832 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.596795 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.597019 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-5nfvz" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.597142 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.598719 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.604119 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8"] Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.756604 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.756659 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/657f3e8c-aa36-488b-9402-dc421f14b392-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.756700 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.756735 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xwqr\" (UniqueName: \"kubernetes.io/projected/657f3e8c-aa36-488b-9402-dc421f14b392-kube-api-access-7xwqr\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.756810 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/657f3e8c-aa36-488b-9402-dc421f14b392-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.857828 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/657f3e8c-aa36-488b-9402-dc421f14b392-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.857905 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.857936 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xwqr\" (UniqueName: \"kubernetes.io/projected/657f3e8c-aa36-488b-9402-dc421f14b392-kube-api-access-7xwqr\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.858020 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/657f3e8c-aa36-488b-9402-dc421f14b392-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.858052 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.860985 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/657f3e8c-aa36-488b-9402-dc421f14b392-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.870030 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.885593 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/657f3e8c-aa36-488b-9402-dc421f14b392-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:29 crc kubenswrapper[4899]: E0123 09:11:29.886023 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 23 09:11:29 crc kubenswrapper[4899]: E0123 09:11:29.886096 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls podName:657f3e8c-aa36-488b-9402-dc421f14b392 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:30.386070493 +0000 UTC m=+1177.253040880 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" (UID: "657f3e8c-aa36-488b-9402-dc421f14b392") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 23 09:11:29 crc kubenswrapper[4899]: I0123 09:11:29.929554 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xwqr\" (UniqueName: \"kubernetes.io/projected/657f3e8c-aa36-488b-9402-dc421f14b392-kube-api-access-7xwqr\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:30 crc kubenswrapper[4899]: I0123 09:11:30.467716 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:30 crc kubenswrapper[4899]: E0123 09:11:30.467927 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 23 09:11:30 crc kubenswrapper[4899]: E0123 09:11:30.467980 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls podName:657f3e8c-aa36-488b-9402-dc421f14b392 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:31.467963529 +0000 UTC m=+1178.334933906 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" (UID: "657f3e8c-aa36-488b-9402-dc421f14b392") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 23 09:11:31 crc kubenswrapper[4899]: I0123 09:11:31.481303 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:31 crc kubenswrapper[4899]: I0123 09:11:31.486666 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/657f3e8c-aa36-488b-9402-dc421f14b392-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8\" (UID: \"657f3e8c-aa36-488b-9402-dc421f14b392\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:31 crc kubenswrapper[4899]: I0123 09:11:31.685880 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:11:31 crc kubenswrapper[4899]: I0123 09:11:31.686197 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:11:31 crc kubenswrapper[4899]: I0123 09:11:31.727150 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" Jan 23 09:11:32 crc kubenswrapper[4899]: I0123 09:11:32.917296 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k"] Jan 23 09:11:32 crc kubenswrapper[4899]: I0123 09:11:32.918765 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:32 crc kubenswrapper[4899]: I0123 09:11:32.921667 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Jan 23 09:11:32 crc kubenswrapper[4899]: I0123 09:11:32.921907 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Jan 23 09:11:32 crc kubenswrapper[4899]: I0123 09:11:32.971819 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k"] Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.015304 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65tf5\" (UniqueName: \"kubernetes.io/projected/914d136a-8ad8-470f-b0cc-19ce201e4217-kube-api-access-65tf5\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.015470 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/914d136a-8ad8-470f-b0cc-19ce201e4217-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.015543 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.015711 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/914d136a-8ad8-470f-b0cc-19ce201e4217-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.015768 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.116664 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/914d136a-8ad8-470f-b0cc-19ce201e4217-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.116713 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.116791 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65tf5\" (UniqueName: \"kubernetes.io/projected/914d136a-8ad8-470f-b0cc-19ce201e4217-kube-api-access-65tf5\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.116820 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/914d136a-8ad8-470f-b0cc-19ce201e4217-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.116846 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: E0123 09:11:33.116979 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 23 09:11:33 crc kubenswrapper[4899]: E0123 09:11:33.117031 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls podName:914d136a-8ad8-470f-b0cc-19ce201e4217 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:33.617014907 +0000 UTC m=+1180.483985294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" (UID: "914d136a-8ad8-470f-b0cc-19ce201e4217") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.117350 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/914d136a-8ad8-470f-b0cc-19ce201e4217-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.118165 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/914d136a-8ad8-470f-b0cc-19ce201e4217-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.125035 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.134401 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65tf5\" (UniqueName: \"kubernetes.io/projected/914d136a-8ad8-470f-b0cc-19ce201e4217-kube-api-access-65tf5\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: I0123 09:11:33.624318 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:33 crc kubenswrapper[4899]: E0123 09:11:33.624700 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 23 09:11:33 crc kubenswrapper[4899]: E0123 09:11:33.624861 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls podName:914d136a-8ad8-470f-b0cc-19ce201e4217 nodeName:}" failed. No retries permitted until 2026-01-23 09:11:34.624841972 +0000 UTC m=+1181.491812359 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" (UID: "914d136a-8ad8-470f-b0cc-19ce201e4217") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 23 09:11:34 crc kubenswrapper[4899]: I0123 09:11:34.640054 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:34 crc kubenswrapper[4899]: I0123 09:11:34.646568 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/914d136a-8ad8-470f-b0cc-19ce201e4217-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k\" (UID: \"914d136a-8ad8-470f-b0cc-19ce201e4217\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:34 crc kubenswrapper[4899]: I0123 09:11:34.738463 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" Jan 23 09:11:35 crc kubenswrapper[4899]: I0123 09:11:35.505077 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8"] Jan 23 09:11:36 crc kubenswrapper[4899]: W0123 09:11:36.435696 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod657f3e8c_aa36_488b_9402_dc421f14b392.slice/crio-27e855d8a13e166c1660111df1d06730fe87961ac468bdc72b148d8541e06555 WatchSource:0}: Error finding container 27e855d8a13e166c1660111df1d06730fe87961ac468bdc72b148d8541e06555: Status 404 returned error can't find the container with id 27e855d8a13e166c1660111df1d06730fe87961ac468bdc72b148d8541e06555 Jan 23 09:11:36 crc kubenswrapper[4899]: I0123 09:11:36.952999 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k"] Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.057563 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"27e855d8a13e166c1660111df1d06730fe87961ac468bdc72b148d8541e06555"} Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.692967 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5"] Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.694348 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.696361 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.697315 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.703571 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5"] Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.827488 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.827541 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/98d82f9c-d552-48b7-9612-70dfd75f8f1a-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.827579 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/98d82f9c-d552-48b7-9612-70dfd75f8f1a-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.827663 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjf5f\" (UniqueName: \"kubernetes.io/projected/98d82f9c-d552-48b7-9612-70dfd75f8f1a-kube-api-access-qjf5f\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.827710 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.928549 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjf5f\" (UniqueName: \"kubernetes.io/projected/98d82f9c-d552-48b7-9612-70dfd75f8f1a-kube-api-access-qjf5f\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.928663 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.928701 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.928725 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/98d82f9c-d552-48b7-9612-70dfd75f8f1a-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.928759 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/98d82f9c-d552-48b7-9612-70dfd75f8f1a-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: E0123 09:11:37.928862 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 23 09:11:37 crc kubenswrapper[4899]: E0123 09:11:37.928939 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls podName:98d82f9c-d552-48b7-9612-70dfd75f8f1a nodeName:}" failed. No retries permitted until 2026-01-23 09:11:38.428921106 +0000 UTC m=+1185.295891493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" (UID: "98d82f9c-d552-48b7-9612-70dfd75f8f1a") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.929508 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/98d82f9c-d552-48b7-9612-70dfd75f8f1a-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.929897 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/98d82f9c-d552-48b7-9612-70dfd75f8f1a-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.935163 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:37 crc kubenswrapper[4899]: I0123 09:11:37.948946 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjf5f\" (UniqueName: \"kubernetes.io/projected/98d82f9c-d552-48b7-9612-70dfd75f8f1a-kube-api-access-qjf5f\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:38 crc kubenswrapper[4899]: I0123 09:11:38.065993 4899 generic.go:334] "Generic (PLEG): container finished" podID="ca3cd04b-86ad-4eaf-a48e-1e985eb010b6" containerID="0e8ddc812be64895b77686e00073a30a49a24ef71aca16da97e5dff781bd2d0f" exitCode=0 Jan 23 09:11:38 crc kubenswrapper[4899]: I0123 09:11:38.066058 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerDied","Data":"0e8ddc812be64895b77686e00073a30a49a24ef71aca16da97e5dff781bd2d0f"} Jan 23 09:11:38 crc kubenswrapper[4899]: I0123 09:11:38.522359 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:38 crc kubenswrapper[4899]: E0123 09:11:38.522515 4899 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 23 09:11:38 crc kubenswrapper[4899]: E0123 09:11:38.522602 4899 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls podName:98d82f9c-d552-48b7-9612-70dfd75f8f1a nodeName:}" failed. No retries permitted until 2026-01-23 09:11:39.522585301 +0000 UTC m=+1186.389555678 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" (UID: "98d82f9c-d552-48b7-9612-70dfd75f8f1a") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.102193 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"1dd3287f55e2a01bb36d801629ee4b19056d1d9e48b46fc51a1a706ffda299a9"} Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.106643 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"88d27067-77f9-40fb-970d-df28779d48a2","Type":"ContainerStarted","Data":"612faea97257fb0c3c91a32bdfc5127561663234ebf9ab90743d9453dec4e110"} Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.116967 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"5832637f105a23cbd9368c3212b3e88c712512cb317db6ff03ecdfb36deaa8ea"} Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.598645 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.607938 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/98d82f9c-d552-48b7-9612-70dfd75f8f1a-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5\" (UID: \"98d82f9c-d552-48b7-9612-70dfd75f8f1a\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:39 crc kubenswrapper[4899]: I0123 09:11:39.823396 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" Jan 23 09:11:40 crc kubenswrapper[4899]: I0123 09:11:40.126824 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"f861d1fd204d1a046716ac27fb5569c40140bc0e24db55c7b590f9aab1c1424a"} Jan 23 09:11:40 crc kubenswrapper[4899]: I0123 09:11:40.654813 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=6.046317983 podStartE2EDuration="51.654755591s" podCreationTimestamp="2026-01-23 09:10:49 +0000 UTC" firstStartedPulling="2026-01-23 09:10:52.817844402 +0000 UTC m=+1139.684814789" lastFinishedPulling="2026-01-23 09:11:38.42628201 +0000 UTC m=+1185.293252397" observedRunningTime="2026-01-23 09:11:39.137438136 +0000 UTC m=+1186.004408533" watchObservedRunningTime="2026-01-23 09:11:40.654755591 +0000 UTC m=+1187.521725978" Jan 23 09:11:40 crc kubenswrapper[4899]: I0123 09:11:40.661062 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5"] Jan 23 09:11:42 crc kubenswrapper[4899]: I0123 09:11:42.151559 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"bbce5bb8a03115d1ca11d9a3d1807f37d07d255bbaf9ebcfc7f8baf2f78a5535"} Jan 23 09:11:42 crc kubenswrapper[4899]: I0123 09:11:42.367570 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Jan 23 09:11:46 crc kubenswrapper[4899]: I0123 09:11:46.974300 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7"] Jan 23 09:11:46 crc kubenswrapper[4899]: I0123 09:11:46.979038 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:46 crc kubenswrapper[4899]: I0123 09:11:46.981764 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Jan 23 09:11:46 crc kubenswrapper[4899]: I0123 09:11:46.982050 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Jan 23 09:11:46 crc kubenswrapper[4899]: I0123 09:11:46.993778 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7"] Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.009863 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/8cc0a786-f743-4af2-93b7-fa04073ee2f0-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.009935 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngx4w\" (UniqueName: \"kubernetes.io/projected/8cc0a786-f743-4af2-93b7-fa04073ee2f0-kube-api-access-ngx4w\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.009974 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/8cc0a786-f743-4af2-93b7-fa04073ee2f0-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.010037 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8cc0a786-f743-4af2-93b7-fa04073ee2f0-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.113038 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8cc0a786-f743-4af2-93b7-fa04073ee2f0-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.113142 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/8cc0a786-f743-4af2-93b7-fa04073ee2f0-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.113531 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8cc0a786-f743-4af2-93b7-fa04073ee2f0-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.113818 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/8cc0a786-f743-4af2-93b7-fa04073ee2f0-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.113877 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngx4w\" (UniqueName: \"kubernetes.io/projected/8cc0a786-f743-4af2-93b7-fa04073ee2f0-kube-api-access-ngx4w\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.114219 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/8cc0a786-f743-4af2-93b7-fa04073ee2f0-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.120989 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/8cc0a786-f743-4af2-93b7-fa04073ee2f0-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.135961 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngx4w\" (UniqueName: \"kubernetes.io/projected/8cc0a786-f743-4af2-93b7-fa04073ee2f0-kube-api-access-ngx4w\") pod \"default-cloud1-coll-event-smartgateway-648848df-pcrw7\" (UID: \"8cc0a786-f743-4af2-93b7-fa04073ee2f0\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.302544 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.534967 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826"] Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.535956 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.541362 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.546526 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826"] Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.625543 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm5h5\" (UniqueName: \"kubernetes.io/projected/2c0ede53-4eca-4cb4-bc55-b01f44be122f-kube-api-access-cm5h5\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.625620 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2c0ede53-4eca-4cb4-bc55-b01f44be122f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.625723 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2c0ede53-4eca-4cb4-bc55-b01f44be122f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.625773 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2c0ede53-4eca-4cb4-bc55-b01f44be122f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.727485 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2c0ede53-4eca-4cb4-bc55-b01f44be122f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.728349 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2c0ede53-4eca-4cb4-bc55-b01f44be122f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.728533 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2c0ede53-4eca-4cb4-bc55-b01f44be122f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.728973 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2c0ede53-4eca-4cb4-bc55-b01f44be122f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.729088 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2c0ede53-4eca-4cb4-bc55-b01f44be122f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.729201 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5h5\" (UniqueName: \"kubernetes.io/projected/2c0ede53-4eca-4cb4-bc55-b01f44be122f-kube-api-access-cm5h5\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.732106 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2c0ede53-4eca-4cb4-bc55-b01f44be122f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.751377 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm5h5\" (UniqueName: \"kubernetes.io/projected/2c0ede53-4eca-4cb4-bc55-b01f44be122f-kube-api-access-cm5h5\") pod \"default-cloud1-ceil-event-smartgateway-78b56f798f-ht826\" (UID: \"2c0ede53-4eca-4cb4-bc55-b01f44be122f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:47 crc kubenswrapper[4899]: I0123 09:11:47.856337 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" Jan 23 09:11:52 crc kubenswrapper[4899]: I0123 09:11:52.367562 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Jan 23 09:11:52 crc kubenswrapper[4899]: I0123 09:11:52.406779 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.133242 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="quay.io/prometheus/alertmanager:latest" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.133568 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:alertmanager,Image:quay.io/prometheus/alertmanager:latest,Command:[],Args:[--config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=127.0.0.1:9093 --web.route-prefix=/ --cluster.label=service-telemetry/default --cluster.peer=alertmanager-default-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:mesh-tcp,HostPort:0,ContainerPort:9094,Protocol:TCP,HostIP:,},ContainerPort{Name:mesh-udp,HostPort:0,ContainerPort:9094,Protocol:UDP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{memory: {{209715200 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-volume,ReadOnly:false,MountPath:/etc/alertmanager/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:true,MountPath:/etc/alertmanager/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-assets,ReadOnly:true,MountPath:/etc/alertmanager/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:alertmanager-default-db,ReadOnly:false,MountPath:/alertmanager,SubPath:alertmanager-db,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-default-alertmanager-proxy-tls,ReadOnly:true,MountPath:/etc/alertmanager/secrets/default-alertmanager-proxy-tls,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-default-session-secret,ReadOnly:true,MountPath:/etc/alertmanager/secrets/default-session-secret,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:web-config,ReadOnly:true,MountPath:/etc/alertmanager/web_config/web-config.yaml,SubPath:web-config.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cluster-tls-config,ReadOnly:true,MountPath:/etc/alertmanager/cluster_tls_config/cluster-tls-config.yaml,SubPath:cluster-tls-config.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7zdrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod alertmanager-default-0_service-telemetry(ca3cd04b-86ad-4eaf-a48e-1e985eb010b6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 23 09:11:53 crc kubenswrapper[4899]: I0123 09:11:53.273484 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.749449 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/sg-bridge:latest" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.750021 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:quay.io/infrawatch/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/anycast/ceilometer/cloud1-metering.sample --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65tf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k_service-telemetry(914d136a-8ad8-470f-b0cc-19ce201e4217): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.766008 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/sg-bridge:latest" Jan 23 09:11:53 crc kubenswrapper[4899]: E0123 09:11:53.768108 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:quay.io/infrawatch/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/collectd/cloud1-telemetry --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7xwqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8_service-telemetry(657f3e8c-aa36-488b-9402-dc421f14b392): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:11:54 crc kubenswrapper[4899]: I0123 09:11:54.174030 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826"] Jan 23 09:11:54 crc kubenswrapper[4899]: I0123 09:11:54.238639 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerStarted","Data":"cdb06322faadfd32805f9a75dd26addda577c8e87478d9b97671f1fb297e9820"} Jan 23 09:11:54 crc kubenswrapper[4899]: I0123 09:11:54.241749 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"afa776f3fa7cf0ed4e1115a6312113d9a6db7d0ebb6448adcaf5e092a1513554"} Jan 23 09:11:54 crc kubenswrapper[4899]: I0123 09:11:54.265102 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7"] Jan 23 09:11:55 crc kubenswrapper[4899]: I0123 09:11:55.251525 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerStarted","Data":"0560cf304f39a9f85e48f60d096dc1d22b9ecb6757d224565dbddbe020e6ece0"} Jan 23 09:11:55 crc kubenswrapper[4899]: I0123 09:11:55.253978 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"2a248261d3225fec29d9dac1c5a1daba159ac41012aa947993a056e6b6add63e"} Jan 23 09:11:55 crc kubenswrapper[4899]: I0123 09:11:55.255529 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerStarted","Data":"b5c3965c279401fc80567d7ee52ecd2736367ed2e768db3509565ffeab850f11"} Jan 23 09:11:55 crc kubenswrapper[4899]: I0123 09:11:55.255577 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerStarted","Data":"cc2be712823d5390ba59fc02703dd0fbc86907246c6e78347d4e168d096167ae"} Jan 23 09:11:56 crc kubenswrapper[4899]: I0123 09:11:56.284236 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerStarted","Data":"56ba76bc986a1fc738b24c9603af0126ef5fcc631c6e4d29476a39b5edd6e7fa"} Jan 23 09:12:01 crc kubenswrapper[4899]: I0123 09:12:01.687808 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:12:01 crc kubenswrapper[4899]: I0123 09:12:01.688333 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:12:01 crc kubenswrapper[4899]: I0123 09:12:01.688380 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:12:01 crc kubenswrapper[4899]: I0123 09:12:01.689070 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:12:01 crc kubenswrapper[4899]: I0123 09:12:01.689126 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6" gracePeriod=600 Jan 23 09:12:02 crc kubenswrapper[4899]: I0123 09:12:02.347049 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6" exitCode=0 Jan 23 09:12:02 crc kubenswrapper[4899]: I0123 09:12:02.347092 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6"} Jan 23 09:12:02 crc kubenswrapper[4899]: I0123 09:12:02.347763 4899 scope.go:117] "RemoveContainer" containerID="103d9e7c7731a2d78677cf024f9288e1f3b843148677dbd0d02e8b86e1bfbc41" Jan 23 09:12:02 crc kubenswrapper[4899]: E0123 09:12:02.987312 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" podUID="914d136a-8ad8-470f-b0cc-19ce201e4217" Jan 23 09:12:02 crc kubenswrapper[4899]: E0123 09:12:02.991620 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" podUID="657f3e8c-aa36-488b-9402-dc421f14b392" Jan 23 09:12:03 crc kubenswrapper[4899]: E0123 09:12:03.276821 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"alertmanager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="service-telemetry/alertmanager-default-0" podUID="ca3cd04b-86ad-4eaf-a48e-1e985eb010b6" Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.356318 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"1c913aeab18dc88b8dac8fac541b9083b6d893a2167d7ffbdf5cde9444a45ae9"} Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.357756 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerStarted","Data":"00ed12f643bb0ad03acaf2ce7d499e9d2a37edcf9df6c302b5b026999b7a3c2f"} Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.361830 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"e91ce42fe6824c61183186127cc9ca1ed612f3c6ce5c7f0de5146d4336f457f7"} Jan 23 09:12:03 crc kubenswrapper[4899]: E0123 09:12:03.362997 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/sg-bridge:latest\\\"\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" podUID="657f3e8c-aa36-488b-9402-dc421f14b392" Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.364518 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc"} Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.366709 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"5eadcd496a3783feada71d80f91c616fbf48b0ce975e2330d2776f861bc67257"} Jan 23 09:12:03 crc kubenswrapper[4899]: E0123 09:12:03.369522 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/sg-bridge:latest\\\"\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" podUID="914d136a-8ad8-470f-b0cc-19ce201e4217" Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.371365 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerStarted","Data":"4e690feed279ee080964efe5e3933f4cd428c4c04bf6b74ebf1fbcb294905247"} Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.376360 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerStarted","Data":"65a4c926d5f3dd20ac44a8d67a2542a2e00f994f64a0e774651eba9a382b01f7"} Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.381622 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" podStartSLOduration=16.756075376 podStartE2EDuration="26.381602423s" podCreationTimestamp="2026-01-23 09:11:37 +0000 UTC" firstStartedPulling="2026-01-23 09:11:53.116615966 +0000 UTC m=+1199.983586353" lastFinishedPulling="2026-01-23 09:12:02.742143013 +0000 UTC m=+1209.609113400" observedRunningTime="2026-01-23 09:12:03.375785031 +0000 UTC m=+1210.242755418" watchObservedRunningTime="2026-01-23 09:12:03.381602423 +0000 UTC m=+1210.248572810" Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.400619 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" podStartSLOduration=8.829914786 podStartE2EDuration="17.400597672s" podCreationTimestamp="2026-01-23 09:11:46 +0000 UTC" firstStartedPulling="2026-01-23 09:11:54.266065281 +0000 UTC m=+1201.133035668" lastFinishedPulling="2026-01-23 09:12:02.836748167 +0000 UTC m=+1209.703718554" observedRunningTime="2026-01-23 09:12:03.394371258 +0000 UTC m=+1210.261341645" watchObservedRunningTime="2026-01-23 09:12:03.400597672 +0000 UTC m=+1210.267568059" Jan 23 09:12:03 crc kubenswrapper[4899]: I0123 09:12:03.514382 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" podStartSLOduration=7.934813325 podStartE2EDuration="16.514357478s" podCreationTimestamp="2026-01-23 09:11:47 +0000 UTC" firstStartedPulling="2026-01-23 09:11:54.184518951 +0000 UTC m=+1201.051489338" lastFinishedPulling="2026-01-23 09:12:02.764063104 +0000 UTC m=+1209.631033491" observedRunningTime="2026-01-23 09:12:03.510326836 +0000 UTC m=+1210.377297233" watchObservedRunningTime="2026-01-23 09:12:03.514357478 +0000 UTC m=+1210.381327855" Jan 23 09:12:04 crc kubenswrapper[4899]: I0123 09:12:04.389205 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"ca3cd04b-86ad-4eaf-a48e-1e985eb010b6","Type":"ContainerStarted","Data":"459009d8d60774131d3bce0ce5eafd49bbac937e2cf73a9e47bb83fe50664fbe"} Jan 23 09:12:04 crc kubenswrapper[4899]: I0123 09:12:04.452906 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=35.825909124 podStartE2EDuration="1m1.452885152s" podCreationTimestamp="2026-01-23 09:11:03 +0000 UTC" firstStartedPulling="2026-01-23 09:11:38.389376703 +0000 UTC m=+1185.256347090" lastFinishedPulling="2026-01-23 09:12:04.016352731 +0000 UTC m=+1210.883323118" observedRunningTime="2026-01-23 09:12:04.444662734 +0000 UTC m=+1211.311633131" watchObservedRunningTime="2026-01-23 09:12:04.452885152 +0000 UTC m=+1211.319855539" Jan 23 09:12:04 crc kubenswrapper[4899]: I0123 09:12:04.689867 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:12:04 crc kubenswrapper[4899]: I0123 09:12:04.690190 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" podUID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" containerName="default-interconnect" containerID="cri-o://490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1" gracePeriod=30 Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.166537 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.287941 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288033 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2k6n\" (UniqueName: \"kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288062 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288114 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288141 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288183 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.288250 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials\") pod \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\" (UID: \"16f3e64c-a1c5-411e-93bf-a523b3f32bbd\") " Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.289048 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.295497 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n" (OuterVolumeSpecName: "kube-api-access-r2k6n") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "kube-api-access-r2k6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.295508 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.295543 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.296163 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.297362 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.297464 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "16f3e64c-a1c5-411e-93bf-a523b3f32bbd" (UID: "16f3e64c-a1c5-411e-93bf-a523b3f32bbd"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389353 4899 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389388 4899 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389404 4899 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389427 4899 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389446 4899 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-config\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389461 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2k6n\" (UniqueName: \"kubernetes.io/projected/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-kube-api-access-r2k6n\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.389473 4899 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/16f3e64c-a1c5-411e-93bf-a523b3f32bbd-sasl-users\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.396531 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"666d1fa17bd44a1f2c4532df239cd471541bdcafceb46ca365135406862cf825"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.398952 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"d734d739cfa78e78edcc64fe9ba8322113675f6c3a3f49758849997773555006"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.400313 4899 generic.go:334] "Generic (PLEG): container finished" podID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" containerID="490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1" exitCode=0 Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.400349 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.400385 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" event={"ID":"16f3e64c-a1c5-411e-93bf-a523b3f32bbd","Type":"ContainerDied","Data":"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.400422 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xqn7w" event={"ID":"16f3e64c-a1c5-411e-93bf-a523b3f32bbd","Type":"ContainerDied","Data":"33c465cd7ecedc0020d67f48af501dfd2f24a160d541bad0bbd41e265f8fc548"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.400440 4899 scope.go:117] "RemoveContainer" containerID="490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.402634 4899 generic.go:334] "Generic (PLEG): container finished" podID="2c0ede53-4eca-4cb4-bc55-b01f44be122f" containerID="0560cf304f39a9f85e48f60d096dc1d22b9ecb6757d224565dbddbe020e6ece0" exitCode=0 Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.402742 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerDied","Data":"0560cf304f39a9f85e48f60d096dc1d22b9ecb6757d224565dbddbe020e6ece0"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.404007 4899 scope.go:117] "RemoveContainer" containerID="0560cf304f39a9f85e48f60d096dc1d22b9ecb6757d224565dbddbe020e6ece0" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.414159 4899 generic.go:334] "Generic (PLEG): container finished" podID="98d82f9c-d552-48b7-9612-70dfd75f8f1a" containerID="2a248261d3225fec29d9dac1c5a1daba159ac41012aa947993a056e6b6add63e" exitCode=0 Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.414309 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerDied","Data":"2a248261d3225fec29d9dac1c5a1daba159ac41012aa947993a056e6b6add63e"} Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.414925 4899 scope.go:117] "RemoveContainer" containerID="2a248261d3225fec29d9dac1c5a1daba159ac41012aa947993a056e6b6add63e" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.422030 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" podStartSLOduration=7.9852858300000005 podStartE2EDuration="36.422009928s" podCreationTimestamp="2026-01-23 09:11:29 +0000 UTC" firstStartedPulling="2026-01-23 09:11:36.476844188 +0000 UTC m=+1183.343814575" lastFinishedPulling="2026-01-23 09:12:04.913568286 +0000 UTC m=+1211.780538673" observedRunningTime="2026-01-23 09:12:05.417722389 +0000 UTC m=+1212.284692796" watchObservedRunningTime="2026-01-23 09:12:05.422009928 +0000 UTC m=+1212.288980315" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.433896 4899 scope.go:117] "RemoveContainer" containerID="490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1" Jan 23 09:12:05 crc kubenswrapper[4899]: E0123 09:12:05.445025 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1\": container with ID starting with 490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1 not found: ID does not exist" containerID="490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.445129 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1"} err="failed to get container status \"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1\": rpc error: code = NotFound desc = could not find container \"490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1\": container with ID starting with 490b9f46d87640312730465f59a00b3060a27c3035c9dc6de944cd6ab12324e1 not found: ID does not exist" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.490921 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" podStartSLOduration=7.042559245 podStartE2EDuration="33.490898456s" podCreationTimestamp="2026-01-23 09:11:32 +0000 UTC" firstStartedPulling="2026-01-23 09:11:38.396405439 +0000 UTC m=+1185.263375826" lastFinishedPulling="2026-01-23 09:12:04.84474465 +0000 UTC m=+1211.711715037" observedRunningTime="2026-01-23 09:12:05.482995936 +0000 UTC m=+1212.349966333" watchObservedRunningTime="2026-01-23 09:12:05.490898456 +0000 UTC m=+1212.357868843" Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.508786 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:12:05 crc kubenswrapper[4899]: I0123 09:12:05.526466 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xqn7w"] Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.132597 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-vrzsc"] Jan 23 09:12:06 crc kubenswrapper[4899]: E0123 09:12:06.132888 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" containerName="default-interconnect" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.132904 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" containerName="default-interconnect" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.133044 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" containerName="default-interconnect" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.133532 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.135106 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.136003 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.136239 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.136299 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.136402 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.136491 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.139854 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-xphnz" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.148236 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-vrzsc"] Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.303810 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x96hr\" (UniqueName: \"kubernetes.io/projected/48780155-85af-4bb5-a73f-6025cf037db1-kube-api-access-x96hr\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.303926 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/48780155-85af-4bb5-a73f-6025cf037db1-sasl-config\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.303952 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.303974 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.304024 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-sasl-users\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.304048 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.304064 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404817 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/48780155-85af-4bb5-a73f-6025cf037db1-sasl-config\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404864 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404887 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404937 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-sasl-users\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404974 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.404994 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.405031 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x96hr\" (UniqueName: \"kubernetes.io/projected/48780155-85af-4bb5-a73f-6025cf037db1-kube-api-access-x96hr\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.406598 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/48780155-85af-4bb5-a73f-6025cf037db1-sasl-config\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.411682 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.417161 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.417546 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-sasl-users\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.418776 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.421829 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/48780155-85af-4bb5-a73f-6025cf037db1-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.424666 4899 generic.go:334] "Generic (PLEG): container finished" podID="8cc0a786-f743-4af2-93b7-fa04073ee2f0" containerID="b5c3965c279401fc80567d7ee52ecd2736367ed2e768db3509565ffeab850f11" exitCode=0 Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.424831 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerDied","Data":"b5c3965c279401fc80567d7ee52ecd2736367ed2e768db3509565ffeab850f11"} Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.425529 4899 scope.go:117] "RemoveContainer" containerID="b5c3965c279401fc80567d7ee52ecd2736367ed2e768db3509565ffeab850f11" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.440771 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x96hr\" (UniqueName: \"kubernetes.io/projected/48780155-85af-4bb5-a73f-6025cf037db1-kube-api-access-x96hr\") pod \"default-interconnect-68864d46cb-vrzsc\" (UID: \"48780155-85af-4bb5-a73f-6025cf037db1\") " pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.447519 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" Jan 23 09:12:06 crc kubenswrapper[4899]: I0123 09:12:06.848142 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-vrzsc"] Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.448236 4899 generic.go:334] "Generic (PLEG): container finished" podID="657f3e8c-aa36-488b-9402-dc421f14b392" containerID="666d1fa17bd44a1f2c4532df239cd471541bdcafceb46ca365135406862cf825" exitCode=0 Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.448410 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerDied","Data":"666d1fa17bd44a1f2c4532df239cd471541bdcafceb46ca365135406862cf825"} Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.449400 4899 scope.go:117] "RemoveContainer" containerID="666d1fa17bd44a1f2c4532df239cd471541bdcafceb46ca365135406862cf825" Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.457260 4899 generic.go:334] "Generic (PLEG): container finished" podID="914d136a-8ad8-470f-b0cc-19ce201e4217" containerID="d734d739cfa78e78edcc64fe9ba8322113675f6c3a3f49758849997773555006" exitCode=0 Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.457388 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerDied","Data":"d734d739cfa78e78edcc64fe9ba8322113675f6c3a3f49758849997773555006"} Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.458023 4899 scope.go:117] "RemoveContainer" containerID="d734d739cfa78e78edcc64fe9ba8322113675f6c3a3f49758849997773555006" Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.461000 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" event={"ID":"48780155-85af-4bb5-a73f-6025cf037db1","Type":"ContainerStarted","Data":"41bbaa303208cafc606c0e6c05eeb15cfa79f4024fe7f116cabc058ca94067a0"} Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.461039 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" event={"ID":"48780155-85af-4bb5-a73f-6025cf037db1","Type":"ContainerStarted","Data":"d1be909a6702b931e96b037d9bb1a19537109f95a01344b6a8ee10227ae98529"} Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.505076 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-vrzsc" podStartSLOduration=3.505056369 podStartE2EDuration="3.505056369s" podCreationTimestamp="2026-01-23 09:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 09:12:07.503944188 +0000 UTC m=+1214.370914575" watchObservedRunningTime="2026-01-23 09:12:07.505056369 +0000 UTC m=+1214.372026756" Jan 23 09:12:07 crc kubenswrapper[4899]: I0123 09:12:07.551926 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f3e64c-a1c5-411e-93bf-a523b3f32bbd" path="/var/lib/kubelet/pods/16f3e64c-a1c5-411e-93bf-a523b3f32bbd/volumes" Jan 23 09:12:08 crc kubenswrapper[4899]: I0123 09:12:08.476972 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerStarted","Data":"9b124299d2feeb676ac6afcbdea6b53a5f15ed3ace4de7e2f956c37749f3c880"} Jan 23 09:12:08 crc kubenswrapper[4899]: I0123 09:12:08.479987 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"43a48d32cd935277e024239d5490b04a8b12a85064d11161e2f4e363897d43a2"} Jan 23 09:12:08 crc kubenswrapper[4899]: I0123 09:12:08.482159 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerStarted","Data":"b2deff5faea14e49c46a71443d70273ebd008e4be267fc28f176ae615c186069"} Jan 23 09:12:08 crc kubenswrapper[4899]: I0123 09:12:08.484948 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"f9144e8edb3ca9685ddfafa3483b2bfa11ad707717a93a099b9739b2975e2779"} Jan 23 09:12:08 crc kubenswrapper[4899]: I0123 09:12:08.487044 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"e1802ea851d00b585f17de99598fb0ce6c7f5aec3e715f8e6d25f76d7e191e36"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.496171 4899 generic.go:334] "Generic (PLEG): container finished" podID="2c0ede53-4eca-4cb4-bc55-b01f44be122f" containerID="9b124299d2feeb676ac6afcbdea6b53a5f15ed3ace4de7e2f956c37749f3c880" exitCode=0 Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.496262 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerDied","Data":"9b124299d2feeb676ac6afcbdea6b53a5f15ed3ace4de7e2f956c37749f3c880"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.496557 4899 scope.go:117] "RemoveContainer" containerID="0560cf304f39a9f85e48f60d096dc1d22b9ecb6757d224565dbddbe020e6ece0" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.497584 4899 scope.go:117] "RemoveContainer" containerID="9b124299d2feeb676ac6afcbdea6b53a5f15ed3ace4de7e2f956c37749f3c880" Jan 23 09:12:09 crc kubenswrapper[4899]: E0123 09:12:09.498502 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-78b56f798f-ht826_service-telemetry(2c0ede53-4eca-4cb4-bc55-b01f44be122f)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" podUID="2c0ede53-4eca-4cb4-bc55-b01f44be122f" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.504300 4899 generic.go:334] "Generic (PLEG): container finished" podID="98d82f9c-d552-48b7-9612-70dfd75f8f1a" containerID="43a48d32cd935277e024239d5490b04a8b12a85064d11161e2f4e363897d43a2" exitCode=0 Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.504375 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerDied","Data":"43a48d32cd935277e024239d5490b04a8b12a85064d11161e2f4e363897d43a2"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.504935 4899 scope.go:117] "RemoveContainer" containerID="43a48d32cd935277e024239d5490b04a8b12a85064d11161e2f4e363897d43a2" Jan 23 09:12:09 crc kubenswrapper[4899]: E0123 09:12:09.505163 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5_service-telemetry(98d82f9c-d552-48b7-9612-70dfd75f8f1a)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" podUID="98d82f9c-d552-48b7-9612-70dfd75f8f1a" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.507405 4899 generic.go:334] "Generic (PLEG): container finished" podID="8cc0a786-f743-4af2-93b7-fa04073ee2f0" containerID="b2deff5faea14e49c46a71443d70273ebd008e4be267fc28f176ae615c186069" exitCode=0 Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.507463 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerDied","Data":"b2deff5faea14e49c46a71443d70273ebd008e4be267fc28f176ae615c186069"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.507923 4899 scope.go:117] "RemoveContainer" containerID="b2deff5faea14e49c46a71443d70273ebd008e4be267fc28f176ae615c186069" Jan 23 09:12:09 crc kubenswrapper[4899]: E0123 09:12:09.508160 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-648848df-pcrw7_service-telemetry(8cc0a786-f743-4af2-93b7-fa04073ee2f0)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" podUID="8cc0a786-f743-4af2-93b7-fa04073ee2f0" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.510818 4899 generic.go:334] "Generic (PLEG): container finished" podID="657f3e8c-aa36-488b-9402-dc421f14b392" containerID="f9144e8edb3ca9685ddfafa3483b2bfa11ad707717a93a099b9739b2975e2779" exitCode=0 Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.510921 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerDied","Data":"f9144e8edb3ca9685ddfafa3483b2bfa11ad707717a93a099b9739b2975e2779"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.511635 4899 scope.go:117] "RemoveContainer" containerID="f9144e8edb3ca9685ddfafa3483b2bfa11ad707717a93a099b9739b2975e2779" Jan 23 09:12:09 crc kubenswrapper[4899]: E0123 09:12:09.511868 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8_service-telemetry(657f3e8c-aa36-488b-9402-dc421f14b392)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" podUID="657f3e8c-aa36-488b-9402-dc421f14b392" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.530637 4899 generic.go:334] "Generic (PLEG): container finished" podID="914d136a-8ad8-470f-b0cc-19ce201e4217" containerID="e1802ea851d00b585f17de99598fb0ce6c7f5aec3e715f8e6d25f76d7e191e36" exitCode=0 Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.531961 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerDied","Data":"e1802ea851d00b585f17de99598fb0ce6c7f5aec3e715f8e6d25f76d7e191e36"} Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.532546 4899 scope.go:117] "RemoveContainer" containerID="e1802ea851d00b585f17de99598fb0ce6c7f5aec3e715f8e6d25f76d7e191e36" Jan 23 09:12:09 crc kubenswrapper[4899]: E0123 09:12:09.532756 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k_service-telemetry(914d136a-8ad8-470f-b0cc-19ce201e4217)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" podUID="914d136a-8ad8-470f-b0cc-19ce201e4217" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.547406 4899 scope.go:117] "RemoveContainer" containerID="2a248261d3225fec29d9dac1c5a1daba159ac41012aa947993a056e6b6add63e" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.605393 4899 scope.go:117] "RemoveContainer" containerID="b5c3965c279401fc80567d7ee52ecd2736367ed2e768db3509565ffeab850f11" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.658230 4899 scope.go:117] "RemoveContainer" containerID="666d1fa17bd44a1f2c4532df239cd471541bdcafceb46ca365135406862cf825" Jan 23 09:12:09 crc kubenswrapper[4899]: I0123 09:12:09.700501 4899 scope.go:117] "RemoveContainer" containerID="d734d739cfa78e78edcc64fe9ba8322113675f6c3a3f49758849997773555006" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.792780 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.795371 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.797967 4899 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.799464 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.805698 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.940351 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-qdr-test-config\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.940420 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:17 crc kubenswrapper[4899]: I0123 09:12:17.940516 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nk2w\" (UniqueName: \"kubernetes.io/projected/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-kube-api-access-7nk2w\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.041804 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-qdr-test-config\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.041899 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.041928 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nk2w\" (UniqueName: \"kubernetes.io/projected/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-kube-api-access-7nk2w\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.042613 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-qdr-test-config\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.055349 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.059784 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nk2w\" (UniqueName: \"kubernetes.io/projected/af2c5766-ae4d-4ff1-8ecb-e1200578fff7-kube-api-access-7nk2w\") pod \"qdr-test\" (UID: \"af2c5766-ae4d-4ff1-8ecb-e1200578fff7\") " pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.121543 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.392271 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 23 09:12:18 crc kubenswrapper[4899]: W0123 09:12:18.393296 4899 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf2c5766_ae4d_4ff1_8ecb_e1200578fff7.slice/crio-714b46d0696066ad350e18797e827cbbf28e5f44242b2425699c209652dd11db WatchSource:0}: Error finding container 714b46d0696066ad350e18797e827cbbf28e5f44242b2425699c209652dd11db: Status 404 returned error can't find the container with id 714b46d0696066ad350e18797e827cbbf28e5f44242b2425699c209652dd11db Jan 23 09:12:18 crc kubenswrapper[4899]: I0123 09:12:18.648879 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"af2c5766-ae4d-4ff1-8ecb-e1200578fff7","Type":"ContainerStarted","Data":"714b46d0696066ad350e18797e827cbbf28e5f44242b2425699c209652dd11db"} Jan 23 09:12:21 crc kubenswrapper[4899]: I0123 09:12:21.514948 4899 scope.go:117] "RemoveContainer" containerID="f9144e8edb3ca9685ddfafa3483b2bfa11ad707717a93a099b9739b2975e2779" Jan 23 09:12:21 crc kubenswrapper[4899]: I0123 09:12:21.515547 4899 scope.go:117] "RemoveContainer" containerID="9b124299d2feeb676ac6afcbdea6b53a5f15ed3ace4de7e2f956c37749f3c880" Jan 23 09:12:21 crc kubenswrapper[4899]: I0123 09:12:21.515647 4899 scope.go:117] "RemoveContainer" containerID="43a48d32cd935277e024239d5490b04a8b12a85064d11161e2f4e363897d43a2" Jan 23 09:12:22 crc kubenswrapper[4899]: I0123 09:12:22.514838 4899 scope.go:117] "RemoveContainer" containerID="b2deff5faea14e49c46a71443d70273ebd008e4be267fc28f176ae615c186069" Jan 23 09:12:24 crc kubenswrapper[4899]: I0123 09:12:24.514467 4899 scope.go:117] "RemoveContainer" containerID="e1802ea851d00b585f17de99598fb0ce6c7f5aec3e715f8e6d25f76d7e191e36" Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.745038 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8" event={"ID":"657f3e8c-aa36-488b-9402-dc421f14b392","Type":"ContainerStarted","Data":"8afdd790e2749352fc87c13564f8ba41afd193d33d8b54f800f41fe19a2955dc"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.747686 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"af2c5766-ae4d-4ff1-8ecb-e1200578fff7","Type":"ContainerStarted","Data":"8587dbcb2a15ad92ecf8d7ec22e0fa8c44ac9f03b16b51588301842d58eac4ac"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.750482 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k" event={"ID":"914d136a-8ad8-470f-b0cc-19ce201e4217","Type":"ContainerStarted","Data":"b3ce482f12c601fd589cb4477cbf73139aa354876245b1cd8a836b4c4386df2d"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.752597 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-78b56f798f-ht826" event={"ID":"2c0ede53-4eca-4cb4-bc55-b01f44be122f","Type":"ContainerStarted","Data":"ba1dcb42619b28754af24786c61beef8ec3e42798cb2a965d23f5468383501f7"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.755253 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5" event={"ID":"98d82f9c-d552-48b7-9612-70dfd75f8f1a","Type":"ContainerStarted","Data":"615a792497e78af7e0f58851cdc90082f71517b7e5d5676b9c4aa8938f574484"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.758000 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-648848df-pcrw7" event={"ID":"8cc0a786-f743-4af2-93b7-fa04073ee2f0","Type":"ContainerStarted","Data":"5b4f6987bc29dde5ddfcb3138df92ff9a1ace74bfbebb10aa9e9a1a088011127"} Jan 23 09:12:28 crc kubenswrapper[4899]: I0123 09:12:28.858367 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.389671139 podStartE2EDuration="11.858349179s" podCreationTimestamp="2026-01-23 09:12:17 +0000 UTC" firstStartedPulling="2026-01-23 09:12:18.395275829 +0000 UTC m=+1225.262246216" lastFinishedPulling="2026-01-23 09:12:27.863953849 +0000 UTC m=+1234.730924256" observedRunningTime="2026-01-23 09:12:28.857204627 +0000 UTC m=+1235.724175014" watchObservedRunningTime="2026-01-23 09:12:28.858349179 +0000 UTC m=+1235.725319566" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.174827 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-2ppq5"] Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.176351 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.182030 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.182386 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.182547 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.182851 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.183102 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.184980 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.204091 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-2ppq5"] Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313097 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313161 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313195 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313265 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313347 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313385 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crj9\" (UniqueName: \"kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.313443 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.415983 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.416776 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.416928 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.417137 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.417419 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.417602 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crj9\" (UniqueName: \"kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.417787 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.417904 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.416971 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.418025 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.419477 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.420912 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.423585 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.441197 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crj9\" (UniqueName: \"kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9\") pod \"stf-smoketest-smoke1-2ppq5\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.501458 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.650905 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.653518 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.654056 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.731264 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc7xl\" (UniqueName: \"kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl\") pod \"curl\" (UID: \"87b06d64-9d08-4533-a96b-936619b5f357\") " pod="service-telemetry/curl" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.832793 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc7xl\" (UniqueName: \"kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl\") pod \"curl\" (UID: \"87b06d64-9d08-4533-a96b-936619b5f357\") " pod="service-telemetry/curl" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.853251 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc7xl\" (UniqueName: \"kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl\") pod \"curl\" (UID: \"87b06d64-9d08-4533-a96b-936619b5f357\") " pod="service-telemetry/curl" Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.962208 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-2ppq5"] Jan 23 09:12:29 crc kubenswrapper[4899]: I0123 09:12:29.986115 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 23 09:12:30 crc kubenswrapper[4899]: I0123 09:12:30.775455 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerStarted","Data":"c3464576ca56edc8e45678fa533d4c4f40596dda4174e6c08ed58bd812007b04"} Jan 23 09:12:33 crc kubenswrapper[4899]: I0123 09:12:33.696809 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 23 09:12:33 crc kubenswrapper[4899]: I0123 09:12:33.794423 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"87b06d64-9d08-4533-a96b-936619b5f357","Type":"ContainerStarted","Data":"ffed4b93e9d2db0d459b0e395e3c54134635d6b61fc387fa1c35fe25cf0f3cfe"} Jan 23 09:12:36 crc kubenswrapper[4899]: I0123 09:12:36.837687 4899 generic.go:334] "Generic (PLEG): container finished" podID="87b06d64-9d08-4533-a96b-936619b5f357" containerID="8478be77d915b7794cb51be2995ec69ef07ddc2a18c6daee09a2faf824df6614" exitCode=0 Jan 23 09:12:36 crc kubenswrapper[4899]: I0123 09:12:36.838295 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"87b06d64-9d08-4533-a96b-936619b5f357","Type":"ContainerDied","Data":"8478be77d915b7794cb51be2995ec69ef07ddc2a18c6daee09a2faf824df6614"} Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.308652 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.436922 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc7xl\" (UniqueName: \"kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl\") pod \"87b06d64-9d08-4533-a96b-936619b5f357\" (UID: \"87b06d64-9d08-4533-a96b-936619b5f357\") " Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.458418 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl" (OuterVolumeSpecName: "kube-api-access-xc7xl") pod "87b06d64-9d08-4533-a96b-936619b5f357" (UID: "87b06d64-9d08-4533-a96b-936619b5f357"). InnerVolumeSpecName "kube-api-access-xc7xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.473198 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_87b06d64-9d08-4533-a96b-936619b5f357/curl/0.log" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.539114 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc7xl\" (UniqueName: \"kubernetes.io/projected/87b06d64-9d08-4533-a96b-936619b5f357-kube-api-access-xc7xl\") on node \"crc\" DevicePath \"\"" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.773813 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-4n76q_930899a8-bdeb-4546-a142-370356b282bd/prometheus-webhook-snmp/0.log" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.973770 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"87b06d64-9d08-4533-a96b-936619b5f357","Type":"ContainerDied","Data":"ffed4b93e9d2db0d459b0e395e3c54134635d6b61fc387fa1c35fe25cf0f3cfe"} Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.973817 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffed4b93e9d2db0d459b0e395e3c54134635d6b61fc387fa1c35fe25cf0f3cfe" Jan 23 09:12:42 crc kubenswrapper[4899]: I0123 09:12:42.973852 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 23 09:12:47 crc kubenswrapper[4899]: I0123 09:12:47.043952 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerStarted","Data":"ae70453e0561095d67bfa4e9bde8d36b8abcc43946974c08abd9ffd59037bb5a"} Jan 23 09:13:01 crc kubenswrapper[4899]: E0123 09:13:01.691390 4899 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo" Jan 23 09:13:01 crc kubenswrapper[4899]: E0123 09:13:01.692036 4899 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:smoketest-ceilometer,Image:quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo,Command:[/smoketest_ceilometer_entrypoint.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CLOUDNAME,Value:smoke1,ValueFrom:nil,},EnvVar{Name:ELASTICSEARCH_AUTH_PASS,Value:lMmV8uggCR3fZun4Lh9Y1Pk2,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_AUTH_TOKEN,Value:eyJhbGciOiJSUzI1NiIsImtpZCI6InF6SnFxNFFjbVk5VmJQZ2dNMmUxdHFmTlJlVWx4UDhSTlhIamV3RUx4WU0ifQ.eyJhdWQiOlsiaHR0cHM6Ly9rdWJlcm5ldGVzLmRlZmF1bHQuc3ZjIl0sImV4cCI6MTc2OTE2MzEzMiwiaWF0IjoxNzY5MTU5NTMyLCJpc3MiOiJodHRwczovL2t1YmVybmV0ZXMuZGVmYXVsdC5zdmMiLCJqdGkiOiJmMmI3ODgzZC1hZDkyLTQ1MmEtYjg2ZC01YTU0NWYyZmMzNmUiLCJrdWJlcm5ldGVzLmlvIjp7Im5hbWVzcGFjZSI6InNlcnZpY2UtdGVsZW1ldHJ5Iiwic2VydmljZWFjY291bnQiOnsibmFtZSI6InN0Zi1wcm9tZXRoZXVzLXJlYWRlciIsInVpZCI6IjY3NGZkNjYzLTZmNTItNDRhOC1iNDE2LTU0ZDNmZjlhZDc4MSJ9fSwibmJmIjoxNzY5MTU5NTMyLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6c2VydmljZS10ZWxlbWV0cnk6c3RmLXByb21ldGhldXMtcmVhZGVyIn0.qTWzfo6OLZVXi5O5oSC1kEAMH6KkupU7-kcTm5TVsmBR1i9CV50YboqjSXP8WN07mubixorPc6MPgZfwX0-sFTkYfniXXr-plzSNnoJhUsqMo1ZAIl9EYqLRO_67Zk3xTgQnOsuZmY2PQ9ySSGE_bdGsTxbpsemwKfFOEOlLuuSXejmsheHAKruH7tlYsXnjlJvDLDJiQulrDI7dBxVflZzLh4L3mpSxuxncdcN5Kwj6T0xU5H3S1XPot-cuC0SuCR4RMWBmztFcEtKBZVBDdIKGpcMAD4JKEvGW40AWnNTsictfxgMVQlHUsNxkxY5GGvTPAugKwdd3MUfmD6Sz6L1hrbzMhoSnSjZH4bvJ5jol2xFPvfP4vpfTqvbQvApZK6CSzPTxf12h3354X5BQc2PR9qLGEoUd4_wQNrwT38AvmCNMt1Z8hMd0JXZ8GHwSvYZKOeAcjh4c7M5O7_nKM_v1uSIZ9OOfi2-FaQyVSwac5vfTHt7wwPuKW-WifXxbyM3SmiHiw5nArXOMCSN_4KZQ8gugYeMavXE4zTaZmi5OYZB0-q7JQYsNX3kU1xbUSJUOKeI0EYp9xhdudnZKdYRM-cCvoYuDilDYCFUTMCa65_vp9tuu8y0AI1r9lYPVi8Zewce7udO5Q4ZMUQR6385HO9AkcP_16-hKN5MNFoE,ValueFrom:nil,},EnvVar{Name:OBSERVABILITY_STRATEGY,Value:<>,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ceilometer-publisher,ReadOnly:false,MountPath:/ceilometer_publish.py,SubPath:ceilometer_publish.py,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceilometer-entrypoint-script,ReadOnly:false,MountPath:/smoketest_ceilometer_entrypoint.sh,SubPath:smoketest_ceilometer_entrypoint.sh,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2crj9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod stf-smoketest-smoke1-2ppq5_service-telemetry(c70be73e-3bbf-4fa1-b4d4-dff666c21826): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 23 09:13:01 crc kubenswrapper[4899]: E0123 09:13:01.693281 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"smoketest-ceilometer\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" Jan 23 09:13:02 crc kubenswrapper[4899]: E0123 09:13:02.393502 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"smoketest-ceilometer\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo\\\"\"" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" Jan 23 09:13:12 crc kubenswrapper[4899]: I0123 09:13:12.936236 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-4n76q_930899a8-bdeb-4546-a142-370356b282bd/prometheus-webhook-snmp/0.log" Jan 23 09:13:15 crc kubenswrapper[4899]: I0123 09:13:15.479192 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerStarted","Data":"66e36b07b1f0088d51d61e5ab70e3753875e06117423634f3a62b213932d1fb1"} Jan 23 09:13:15 crc kubenswrapper[4899]: I0123 09:13:15.494603 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" podStartSLOduration=1.486928311 podStartE2EDuration="46.494587468s" podCreationTimestamp="2026-01-23 09:12:29 +0000 UTC" firstStartedPulling="2026-01-23 09:12:29.976465942 +0000 UTC m=+1236.843436329" lastFinishedPulling="2026-01-23 09:13:14.984125099 +0000 UTC m=+1281.851095486" observedRunningTime="2026-01-23 09:13:15.493274812 +0000 UTC m=+1282.360245199" watchObservedRunningTime="2026-01-23 09:13:15.494587468 +0000 UTC m=+1282.361557855" Jan 23 09:13:21 crc kubenswrapper[4899]: I0123 09:13:21.517003 4899 generic.go:334] "Generic (PLEG): container finished" podID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerID="ae70453e0561095d67bfa4e9bde8d36b8abcc43946974c08abd9ffd59037bb5a" exitCode=0 Jan 23 09:13:21 crc kubenswrapper[4899]: I0123 09:13:21.525467 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerDied","Data":"ae70453e0561095d67bfa4e9bde8d36b8abcc43946974c08abd9ffd59037bb5a"} Jan 23 09:13:21 crc kubenswrapper[4899]: I0123 09:13:21.526078 4899 scope.go:117] "RemoveContainer" containerID="ae70453e0561095d67bfa4e9bde8d36b8abcc43946974c08abd9ffd59037bb5a" Jan 23 09:13:47 crc kubenswrapper[4899]: I0123 09:13:47.704807 4899 generic.go:334] "Generic (PLEG): container finished" podID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerID="66e36b07b1f0088d51d61e5ab70e3753875e06117423634f3a62b213932d1fb1" exitCode=0 Jan 23 09:13:47 crc kubenswrapper[4899]: I0123 09:13:47.704887 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerDied","Data":"66e36b07b1f0088d51d61e5ab70e3753875e06117423634f3a62b213932d1fb1"} Jan 23 09:13:48 crc kubenswrapper[4899]: I0123 09:13:48.954312 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105089 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crj9\" (UniqueName: \"kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105255 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105300 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105329 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105398 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105428 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.105458 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script\") pod \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\" (UID: \"c70be73e-3bbf-4fa1-b4d4-dff666c21826\") " Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.114358 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9" (OuterVolumeSpecName: "kube-api-access-2crj9") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "kube-api-access-2crj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.122413 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.122558 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.123117 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.128972 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.130733 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.132771 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "c70be73e-3bbf-4fa1-b4d4-dff666c21826" (UID: "c70be73e-3bbf-4fa1-b4d4-dff666c21826"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207533 4899 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-config\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207576 4899 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-sensubility-config\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207590 4899 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-healthcheck-log\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207601 4899 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207615 4899 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207627 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crj9\" (UniqueName: \"kubernetes.io/projected/c70be73e-3bbf-4fa1-b4d4-dff666c21826-kube-api-access-2crj9\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.207640 4899 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c70be73e-3bbf-4fa1-b4d4-dff666c21826-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.722050 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" event={"ID":"c70be73e-3bbf-4fa1-b4d4-dff666c21826","Type":"ContainerDied","Data":"c3464576ca56edc8e45678fa533d4c4f40596dda4174e6c08ed58bd812007b04"} Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.722350 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3464576ca56edc8e45678fa533d4c4f40596dda4174e6c08ed58bd812007b04" Jan 23 09:13:49 crc kubenswrapper[4899]: I0123 09:13:49.722197 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-2ppq5" Jan 23 09:13:50 crc kubenswrapper[4899]: I0123 09:13:50.933053 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-2ppq5_c70be73e-3bbf-4fa1-b4d4-dff666c21826/smoketest-collectd/0.log" Jan 23 09:13:51 crc kubenswrapper[4899]: I0123 09:13:51.297413 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-2ppq5_c70be73e-3bbf-4fa1-b4d4-dff666c21826/smoketest-ceilometer/0.log" Jan 23 09:13:51 crc kubenswrapper[4899]: I0123 09:13:51.558393 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-vrzsc_48780155-85af-4bb5-a73f-6025cf037db1/default-interconnect/0.log" Jan 23 09:13:51 crc kubenswrapper[4899]: I0123 09:13:51.815683 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8_657f3e8c-aa36-488b-9402-dc421f14b392/bridge/2.log" Jan 23 09:13:52 crc kubenswrapper[4899]: I0123 09:13:52.067030 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-w2lc8_657f3e8c-aa36-488b-9402-dc421f14b392/sg-core/0.log" Jan 23 09:13:52 crc kubenswrapper[4899]: I0123 09:13:52.405702 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-648848df-pcrw7_8cc0a786-f743-4af2-93b7-fa04073ee2f0/bridge/2.log" Jan 23 09:13:52 crc kubenswrapper[4899]: I0123 09:13:52.684530 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-648848df-pcrw7_8cc0a786-f743-4af2-93b7-fa04073ee2f0/sg-core/0.log" Jan 23 09:13:52 crc kubenswrapper[4899]: I0123 09:13:52.939833 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k_914d136a-8ad8-470f-b0cc-19ce201e4217/bridge/2.log" Jan 23 09:13:53 crc kubenswrapper[4899]: I0123 09:13:53.225242 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-tzm2k_914d136a-8ad8-470f-b0cc-19ce201e4217/sg-core/0.log" Jan 23 09:13:53 crc kubenswrapper[4899]: I0123 09:13:53.491904 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-78b56f798f-ht826_2c0ede53-4eca-4cb4-bc55-b01f44be122f/bridge/2.log" Jan 23 09:13:53 crc kubenswrapper[4899]: I0123 09:13:53.780751 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-78b56f798f-ht826_2c0ede53-4eca-4cb4-bc55-b01f44be122f/sg-core/0.log" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.039093 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5_98d82f9c-d552-48b7-9612-70dfd75f8f1a/bridge/2.log" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.313283 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-kqrc5_98d82f9c-d552-48b7-9612-70dfd75f8f1a/sg-core/0.log" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481383 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:13:54 crc kubenswrapper[4899]: E0123 09:13:54.481703 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-ceilometer" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481727 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-ceilometer" Jan 23 09:13:54 crc kubenswrapper[4899]: E0123 09:13:54.481751 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-collectd" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481759 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-collectd" Jan 23 09:13:54 crc kubenswrapper[4899]: E0123 09:13:54.481776 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b06d64-9d08-4533-a96b-936619b5f357" containerName="curl" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481784 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b06d64-9d08-4533-a96b-936619b5f357" containerName="curl" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481939 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b06d64-9d08-4533-a96b-936619b5f357" containerName="curl" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481953 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-collectd" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.481975 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70be73e-3bbf-4fa1-b4d4-dff666c21826" containerName="smoketest-ceilometer" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.482530 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.489553 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.577218 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsn56\" (UniqueName: \"kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56\") pod \"infrawatch-operators-hbn6n\" (UID: \"43bf29d0-c34b-4137-b815-170c5bbe3a86\") " pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.678270 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsn56\" (UniqueName: \"kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56\") pod \"infrawatch-operators-hbn6n\" (UID: \"43bf29d0-c34b-4137-b815-170c5bbe3a86\") " pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.700246 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsn56\" (UniqueName: \"kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56\") pod \"infrawatch-operators-hbn6n\" (UID: \"43bf29d0-c34b-4137-b815-170c5bbe3a86\") " pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:13:54 crc kubenswrapper[4899]: I0123 09:13:54.810028 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:13:55 crc kubenswrapper[4899]: I0123 09:13:55.264032 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:13:55 crc kubenswrapper[4899]: I0123 09:13:55.778589 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-hbn6n" event={"ID":"43bf29d0-c34b-4137-b815-170c5bbe3a86","Type":"ContainerStarted","Data":"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215"} Jan 23 09:13:55 crc kubenswrapper[4899]: I0123 09:13:55.778853 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-hbn6n" event={"ID":"43bf29d0-c34b-4137-b815-170c5bbe3a86","Type":"ContainerStarted","Data":"b89ee75f7eb6e862a5e3a43a965552afdae88763a5646f605d17a12c130bbeff"} Jan 23 09:13:55 crc kubenswrapper[4899]: I0123 09:13:55.805809 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-hbn6n" podStartSLOduration=1.6430219080000001 podStartE2EDuration="1.805791145s" podCreationTimestamp="2026-01-23 09:13:54 +0000 UTC" firstStartedPulling="2026-01-23 09:13:55.27878294 +0000 UTC m=+1322.145753327" lastFinishedPulling="2026-01-23 09:13:55.441552177 +0000 UTC m=+1322.308522564" observedRunningTime="2026-01-23 09:13:55.797223882 +0000 UTC m=+1322.664194269" watchObservedRunningTime="2026-01-23 09:13:55.805791145 +0000 UTC m=+1322.672761532" Jan 23 09:13:57 crc kubenswrapper[4899]: I0123 09:13:57.106026 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bbbc889bc-dgksc_566e53d3-281e-4452-a072-d387390d427c/operator/0.log" Jan 23 09:13:57 crc kubenswrapper[4899]: I0123 09:13:57.375601 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_88d27067-77f9-40fb-970d-df28779d48a2/prometheus/0.log" Jan 23 09:13:57 crc kubenswrapper[4899]: I0123 09:13:57.670979 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_bff1497a-cc11-4f67-9f66-3fba05dceef1/elasticsearch/0.log" Jan 23 09:13:57 crc kubenswrapper[4899]: I0123 09:13:57.940496 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-4n76q_930899a8-bdeb-4546-a142-370356b282bd/prometheus-webhook-snmp/0.log" Jan 23 09:13:58 crc kubenswrapper[4899]: I0123 09:13:58.184563 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_ca3cd04b-86ad-4eaf-a48e-1e985eb010b6/alertmanager/0.log" Jan 23 09:14:04 crc kubenswrapper[4899]: I0123 09:14:04.811117 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:04 crc kubenswrapper[4899]: I0123 09:14:04.811793 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:04 crc kubenswrapper[4899]: I0123 09:14:04.838895 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:04 crc kubenswrapper[4899]: I0123 09:14:04.868293 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:05 crc kubenswrapper[4899]: I0123 09:14:05.068961 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:14:06 crc kubenswrapper[4899]: I0123 09:14:06.850023 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-hbn6n" podUID="43bf29d0-c34b-4137-b815-170c5bbe3a86" containerName="registry-server" containerID="cri-o://5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215" gracePeriod=2 Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.202334 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.361480 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsn56\" (UniqueName: \"kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56\") pod \"43bf29d0-c34b-4137-b815-170c5bbe3a86\" (UID: \"43bf29d0-c34b-4137-b815-170c5bbe3a86\") " Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.368027 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56" (OuterVolumeSpecName: "kube-api-access-gsn56") pod "43bf29d0-c34b-4137-b815-170c5bbe3a86" (UID: "43bf29d0-c34b-4137-b815-170c5bbe3a86"). InnerVolumeSpecName "kube-api-access-gsn56". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.464121 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsn56\" (UniqueName: \"kubernetes.io/projected/43bf29d0-c34b-4137-b815-170c5bbe3a86-kube-api-access-gsn56\") on node \"crc\" DevicePath \"\"" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.857759 4899 generic.go:334] "Generic (PLEG): container finished" podID="43bf29d0-c34b-4137-b815-170c5bbe3a86" containerID="5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215" exitCode=0 Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.857795 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-hbn6n" event={"ID":"43bf29d0-c34b-4137-b815-170c5bbe3a86","Type":"ContainerDied","Data":"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215"} Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.857819 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-hbn6n" event={"ID":"43bf29d0-c34b-4137-b815-170c5bbe3a86","Type":"ContainerDied","Data":"b89ee75f7eb6e862a5e3a43a965552afdae88763a5646f605d17a12c130bbeff"} Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.857836 4899 scope.go:117] "RemoveContainer" containerID="5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.857837 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-hbn6n" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.877841 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.880454 4899 scope.go:117] "RemoveContainer" containerID="5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215" Jan 23 09:14:07 crc kubenswrapper[4899]: E0123 09:14:07.880793 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215\": container with ID starting with 5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215 not found: ID does not exist" containerID="5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.880823 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215"} err="failed to get container status \"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215\": rpc error: code = NotFound desc = could not find container \"5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215\": container with ID starting with 5d43cc49dd46acd3ba74774f4766be37eb2bd00bb40a7c34f16f92dad04e3215 not found: ID does not exist" Jan 23 09:14:07 crc kubenswrapper[4899]: I0123 09:14:07.883329 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-hbn6n"] Jan 23 09:14:09 crc kubenswrapper[4899]: I0123 09:14:09.523194 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43bf29d0-c34b-4137-b815-170c5bbe3a86" path="/var/lib/kubelet/pods/43bf29d0-c34b-4137-b815-170c5bbe3a86/volumes" Jan 23 09:14:11 crc kubenswrapper[4899]: I0123 09:14:11.557775 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-55b89ddfb9-jgdbv_1ca0b316-d4d5-4c10-aff5-24bd8fde1d49/operator/0.log" Jan 23 09:14:14 crc kubenswrapper[4899]: I0123 09:14:14.869143 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bbbc889bc-dgksc_566e53d3-281e-4452-a072-d387390d427c/operator/0.log" Jan 23 09:14:15 crc kubenswrapper[4899]: I0123 09:14:15.173512 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_af2c5766-ae4d-4ff1-8ecb-e1200578fff7/qdr/0.log" Jan 23 09:14:31 crc kubenswrapper[4899]: I0123 09:14:31.686133 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:14:31 crc kubenswrapper[4899]: I0123 09:14:31.687393 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.728308 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qsppb/must-gather-m9dh9"] Jan 23 09:14:49 crc kubenswrapper[4899]: E0123 09:14:49.729148 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bf29d0-c34b-4137-b815-170c5bbe3a86" containerName="registry-server" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.729162 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bf29d0-c34b-4137-b815-170c5bbe3a86" containerName="registry-server" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.729344 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="43bf29d0-c34b-4137-b815-170c5bbe3a86" containerName="registry-server" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.730184 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.733699 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qsppb"/"openshift-service-ca.crt" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.733895 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qsppb"/"kube-root-ca.crt" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.735637 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qsppb"/"default-dockercfg-bg2wp" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.755788 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qsppb/must-gather-m9dh9"] Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.840575 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8kq\" (UniqueName: \"kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.840628 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.942816 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8kq\" (UniqueName: \"kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.942857 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.943237 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:49 crc kubenswrapper[4899]: I0123 09:14:49.965803 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8kq\" (UniqueName: \"kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq\") pod \"must-gather-m9dh9\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:50 crc kubenswrapper[4899]: I0123 09:14:50.049888 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:14:50 crc kubenswrapper[4899]: I0123 09:14:50.454142 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qsppb/must-gather-m9dh9"] Jan 23 09:14:51 crc kubenswrapper[4899]: I0123 09:14:51.179947 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qsppb/must-gather-m9dh9" event={"ID":"bfbde8db-491d-4d66-849b-c0ad6abcc433","Type":"ContainerStarted","Data":"5584c4d4e1ff22da84b825e2bfb22d151147ed92c117625d69a78ed4c2bc368b"} Jan 23 09:14:57 crc kubenswrapper[4899]: I0123 09:14:57.991052 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:14:57 crc kubenswrapper[4899]: I0123 09:14:57.993269 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.005853 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.071774 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.072188 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.072335 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bxm\" (UniqueName: \"kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.173182 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.173269 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.173303 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bxm\" (UniqueName: \"kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.174153 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.174387 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.201383 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bxm\" (UniqueName: \"kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm\") pod \"redhat-operators-q7k98\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.246666 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qsppb/must-gather-m9dh9" event={"ID":"bfbde8db-491d-4d66-849b-c0ad6abcc433","Type":"ContainerStarted","Data":"10ab614b799ec1ff38e96f74c6c7c2e7f4fdbe50265ebc57e68921c4441c310e"} Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.313599 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:14:58 crc kubenswrapper[4899]: I0123 09:14:58.568193 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:14:59 crc kubenswrapper[4899]: I0123 09:14:59.254332 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qsppb/must-gather-m9dh9" event={"ID":"bfbde8db-491d-4d66-849b-c0ad6abcc433","Type":"ContainerStarted","Data":"fb4c0a30ee9849139c728da10db1d8a0e813df0438bf9f4155afb80aed2d0f19"} Jan 23 09:14:59 crc kubenswrapper[4899]: I0123 09:14:59.255900 4899 generic.go:334] "Generic (PLEG): container finished" podID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerID="67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627" exitCode=0 Jan 23 09:14:59 crc kubenswrapper[4899]: I0123 09:14:59.255940 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerDied","Data":"67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627"} Jan 23 09:14:59 crc kubenswrapper[4899]: I0123 09:14:59.255966 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerStarted","Data":"ad99d29133c8fd10cd81ce024a5c10fb18b9b756940a49d51b3e89073344e409"} Jan 23 09:14:59 crc kubenswrapper[4899]: I0123 09:14:59.274869 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qsppb/must-gather-m9dh9" podStartSLOduration=2.888144228 podStartE2EDuration="10.274848266s" podCreationTimestamp="2026-01-23 09:14:49 +0000 UTC" firstStartedPulling="2026-01-23 09:14:50.470171398 +0000 UTC m=+1377.337141785" lastFinishedPulling="2026-01-23 09:14:57.856875436 +0000 UTC m=+1384.723845823" observedRunningTime="2026-01-23 09:14:59.271075363 +0000 UTC m=+1386.138045770" watchObservedRunningTime="2026-01-23 09:14:59.274848266 +0000 UTC m=+1386.141818663" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.149248 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn"] Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.150537 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.153654 4899 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.156493 4899 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.172666 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn"] Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.203274 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.203375 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2lr7\" (UniqueName: \"kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.203418 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.274119 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerStarted","Data":"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd"} Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.305073 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.306108 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.306769 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2lr7\" (UniqueName: \"kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.306998 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.317079 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.325224 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2lr7\" (UniqueName: \"kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7\") pod \"collect-profiles-29485995-7jjtn\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.477174 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:00 crc kubenswrapper[4899]: I0123 09:15:00.935065 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn"] Jan 23 09:15:01 crc kubenswrapper[4899]: I0123 09:15:01.283158 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" event={"ID":"b0ed0ab1-d385-4204-85a9-724f2b327425","Type":"ContainerStarted","Data":"ceb5cfde8130ae363c65095840f3991cb3357c8ce99bd6239981b803895f01ef"} Jan 23 09:15:01 crc kubenswrapper[4899]: I0123 09:15:01.686323 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:15:01 crc kubenswrapper[4899]: I0123 09:15:01.686410 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:15:02 crc kubenswrapper[4899]: I0123 09:15:02.292484 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" event={"ID":"b0ed0ab1-d385-4204-85a9-724f2b327425","Type":"ContainerStarted","Data":"9cf5b0a35c789d33bbae1b22467736679d5057213a1d7bdbab6796b00817b1fb"} Jan 23 09:15:02 crc kubenswrapper[4899]: I0123 09:15:02.295557 4899 generic.go:334] "Generic (PLEG): container finished" podID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerID="b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd" exitCode=0 Jan 23 09:15:02 crc kubenswrapper[4899]: I0123 09:15:02.295599 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerDied","Data":"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd"} Jan 23 09:15:02 crc kubenswrapper[4899]: I0123 09:15:02.307334 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" podStartSLOduration=2.307313501 podStartE2EDuration="2.307313501s" podCreationTimestamp="2026-01-23 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-23 09:15:02.305927544 +0000 UTC m=+1389.172897941" watchObservedRunningTime="2026-01-23 09:15:02.307313501 +0000 UTC m=+1389.174283888" Jan 23 09:15:04 crc kubenswrapper[4899]: I0123 09:15:04.310106 4899 generic.go:334] "Generic (PLEG): container finished" podID="b0ed0ab1-d385-4204-85a9-724f2b327425" containerID="9cf5b0a35c789d33bbae1b22467736679d5057213a1d7bdbab6796b00817b1fb" exitCode=0 Jan 23 09:15:04 crc kubenswrapper[4899]: I0123 09:15:04.310178 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" event={"ID":"b0ed0ab1-d385-4204-85a9-724f2b327425","Type":"ContainerDied","Data":"9cf5b0a35c789d33bbae1b22467736679d5057213a1d7bdbab6796b00817b1fb"} Jan 23 09:15:04 crc kubenswrapper[4899]: I0123 09:15:04.313071 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerStarted","Data":"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb"} Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.568896 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.587740 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q7k98" podStartSLOduration=4.434686756 podStartE2EDuration="8.587720435s" podCreationTimestamp="2026-01-23 09:14:57 +0000 UTC" firstStartedPulling="2026-01-23 09:14:59.257725889 +0000 UTC m=+1386.124696276" lastFinishedPulling="2026-01-23 09:15:03.410759568 +0000 UTC m=+1390.277729955" observedRunningTime="2026-01-23 09:15:04.344732526 +0000 UTC m=+1391.211702923" watchObservedRunningTime="2026-01-23 09:15:05.587720435 +0000 UTC m=+1392.454690822" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.691183 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume\") pod \"b0ed0ab1-d385-4204-85a9-724f2b327425\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.691322 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2lr7\" (UniqueName: \"kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7\") pod \"b0ed0ab1-d385-4204-85a9-724f2b327425\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.691361 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume\") pod \"b0ed0ab1-d385-4204-85a9-724f2b327425\" (UID: \"b0ed0ab1-d385-4204-85a9-724f2b327425\") " Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.692009 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume" (OuterVolumeSpecName: "config-volume") pod "b0ed0ab1-d385-4204-85a9-724f2b327425" (UID: "b0ed0ab1-d385-4204-85a9-724f2b327425"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.697722 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b0ed0ab1-d385-4204-85a9-724f2b327425" (UID: "b0ed0ab1-d385-4204-85a9-724f2b327425"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.697812 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7" (OuterVolumeSpecName: "kube-api-access-f2lr7") pod "b0ed0ab1-d385-4204-85a9-724f2b327425" (UID: "b0ed0ab1-d385-4204-85a9-724f2b327425"). InnerVolumeSpecName "kube-api-access-f2lr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.793413 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2lr7\" (UniqueName: \"kubernetes.io/projected/b0ed0ab1-d385-4204-85a9-724f2b327425-kube-api-access-f2lr7\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.793466 4899 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0ed0ab1-d385-4204-85a9-724f2b327425-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:05 crc kubenswrapper[4899]: I0123 09:15:05.793477 4899 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0ed0ab1-d385-4204-85a9-724f2b327425-config-volume\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:06 crc kubenswrapper[4899]: I0123 09:15:06.327411 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" event={"ID":"b0ed0ab1-d385-4204-85a9-724f2b327425","Type":"ContainerDied","Data":"ceb5cfde8130ae363c65095840f3991cb3357c8ce99bd6239981b803895f01ef"} Jan 23 09:15:06 crc kubenswrapper[4899]: I0123 09:15:06.327459 4899 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb5cfde8130ae363c65095840f3991cb3357c8ce99bd6239981b803895f01ef" Jan 23 09:15:06 crc kubenswrapper[4899]: I0123 09:15:06.327473 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29485995-7jjtn" Jan 23 09:15:08 crc kubenswrapper[4899]: I0123 09:15:08.313822 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:08 crc kubenswrapper[4899]: I0123 09:15:08.315386 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:09 crc kubenswrapper[4899]: I0123 09:15:09.356888 4899 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q7k98" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="registry-server" probeResult="failure" output=< Jan 23 09:15:09 crc kubenswrapper[4899]: timeout: failed to connect service ":50051" within 1s Jan 23 09:15:09 crc kubenswrapper[4899]: > Jan 23 09:15:18 crc kubenswrapper[4899]: I0123 09:15:18.361573 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:18 crc kubenswrapper[4899]: I0123 09:15:18.419234 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:18 crc kubenswrapper[4899]: I0123 09:15:18.603612 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.409912 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q7k98" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="registry-server" containerID="cri-o://0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb" gracePeriod=2 Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.760503 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.928127 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5bxm\" (UniqueName: \"kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm\") pod \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.928359 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content\") pod \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.928553 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities\") pod \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\" (UID: \"56ce0e86-a8dd-4076-b39a-3e62ad092ac9\") " Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.929034 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities" (OuterVolumeSpecName: "utilities") pod "56ce0e86-a8dd-4076-b39a-3e62ad092ac9" (UID: "56ce0e86-a8dd-4076-b39a-3e62ad092ac9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:15:19 crc kubenswrapper[4899]: I0123 09:15:19.937774 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm" (OuterVolumeSpecName: "kube-api-access-d5bxm") pod "56ce0e86-a8dd-4076-b39a-3e62ad092ac9" (UID: "56ce0e86-a8dd-4076-b39a-3e62ad092ac9"). InnerVolumeSpecName "kube-api-access-d5bxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.030762 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.030807 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5bxm\" (UniqueName: \"kubernetes.io/projected/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-kube-api-access-d5bxm\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.074425 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56ce0e86-a8dd-4076-b39a-3e62ad092ac9" (UID: "56ce0e86-a8dd-4076-b39a-3e62ad092ac9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.131791 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56ce0e86-a8dd-4076-b39a-3e62ad092ac9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.418496 4899 generic.go:334] "Generic (PLEG): container finished" podID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerID="0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb" exitCode=0 Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.418549 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerDied","Data":"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb"} Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.418584 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q7k98" event={"ID":"56ce0e86-a8dd-4076-b39a-3e62ad092ac9","Type":"ContainerDied","Data":"ad99d29133c8fd10cd81ce024a5c10fb18b9b756940a49d51b3e89073344e409"} Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.418586 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q7k98" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.418606 4899 scope.go:117] "RemoveContainer" containerID="0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.441375 4899 scope.go:117] "RemoveContainer" containerID="b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.466233 4899 scope.go:117] "RemoveContainer" containerID="67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.468148 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.473823 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q7k98"] Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.482793 4899 scope.go:117] "RemoveContainer" containerID="0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb" Jan 23 09:15:20 crc kubenswrapper[4899]: E0123 09:15:20.483273 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb\": container with ID starting with 0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb not found: ID does not exist" containerID="0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.483326 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb"} err="failed to get container status \"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb\": rpc error: code = NotFound desc = could not find container \"0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb\": container with ID starting with 0c53f1cf2f2252ac095e0962c417b1b32ca553cf07491aecfda453a3217807fb not found: ID does not exist" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.483356 4899 scope.go:117] "RemoveContainer" containerID="b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd" Jan 23 09:15:20 crc kubenswrapper[4899]: E0123 09:15:20.483763 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd\": container with ID starting with b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd not found: ID does not exist" containerID="b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.483816 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd"} err="failed to get container status \"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd\": rpc error: code = NotFound desc = could not find container \"b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd\": container with ID starting with b82449101e38d0cae610784e1bf2f6c72a9878cb32d7b14132d155373e369ecd not found: ID does not exist" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.483849 4899 scope.go:117] "RemoveContainer" containerID="67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627" Jan 23 09:15:20 crc kubenswrapper[4899]: E0123 09:15:20.484131 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627\": container with ID starting with 67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627 not found: ID does not exist" containerID="67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627" Jan 23 09:15:20 crc kubenswrapper[4899]: I0123 09:15:20.484160 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627"} err="failed to get container status \"67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627\": rpc error: code = NotFound desc = could not find container \"67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627\": container with ID starting with 67769b31d90986aa2f6dc50897e4b698f4887f930e32f8783e8893e62f124627 not found: ID does not exist" Jan 23 09:15:21 crc kubenswrapper[4899]: I0123 09:15:21.524704 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" path="/var/lib/kubelet/pods/56ce0e86-a8dd-4076-b39a-3e62ad092ac9/volumes" Jan 23 09:15:31 crc kubenswrapper[4899]: I0123 09:15:31.685671 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:15:31 crc kubenswrapper[4899]: I0123 09:15:31.687108 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:15:31 crc kubenswrapper[4899]: I0123 09:15:31.687307 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:15:31 crc kubenswrapper[4899]: I0123 09:15:31.687925 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:15:31 crc kubenswrapper[4899]: I0123 09:15:31.688038 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc" gracePeriod=600 Jan 23 09:15:32 crc kubenswrapper[4899]: I0123 09:15:32.501353 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc" exitCode=0 Jan 23 09:15:32 crc kubenswrapper[4899]: I0123 09:15:32.501473 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc"} Jan 23 09:15:32 crc kubenswrapper[4899]: I0123 09:15:32.501959 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerStarted","Data":"49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3"} Jan 23 09:15:32 crc kubenswrapper[4899]: I0123 09:15:32.502016 4899 scope.go:117] "RemoveContainer" containerID="367fd41bae3e14849e26f97deed61a5f60231d2838de383bafbfadd559c279b6" Jan 23 09:15:41 crc kubenswrapper[4899]: I0123 09:15:41.332974 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qfb94_65baaf43-84aa-4f18-99ab-a7816e4600a4/control-plane-machine-set-operator/0.log" Jan 23 09:15:41 crc kubenswrapper[4899]: I0123 09:15:41.517003 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tbwm2_926c5bb9-7f91-4ea9-860c-694ec9c5c4db/kube-rbac-proxy/0.log" Jan 23 09:15:41 crc kubenswrapper[4899]: I0123 09:15:41.520431 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tbwm2_926c5bb9-7f91-4ea9-860c-694ec9c5c4db/machine-api-operator/0.log" Jan 23 09:15:52 crc kubenswrapper[4899]: I0123 09:15:52.557658 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-zmxfr_067ef4a9-ba7d-4d41-8b04-c3180e74c354/cert-manager-controller/0.log" Jan 23 09:15:52 crc kubenswrapper[4899]: I0123 09:15:52.703534 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-6zpnt_db949b26-d7d6-45e7-abd0-70953e7a713d/cert-manager-cainjector/0.log" Jan 23 09:15:52 crc kubenswrapper[4899]: I0123 09:15:52.765687 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-xbh7b_e0d0ee40-1af4-479a-84f9-87322410c93f/cert-manager-webhook/0.log" Jan 23 09:16:06 crc kubenswrapper[4899]: I0123 09:16:06.139182 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-5dpg7_a9bf5caa-85e0-4b70-8145-d6c25fa15999/prometheus-operator/0.log" Jan 23 09:16:06 crc kubenswrapper[4899]: I0123 09:16:06.293291 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd_ab9f7afb-bc13-43ce-b218-bdb18a103b51/prometheus-operator-admission-webhook/0.log" Jan 23 09:16:06 crc kubenswrapper[4899]: I0123 09:16:06.322824 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw_e90996a1-43d9-44c4-b23e-e8b110764326/prometheus-operator-admission-webhook/0.log" Jan 23 09:16:06 crc kubenswrapper[4899]: I0123 09:16:06.481682 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-g9rtj_6a152f52-cbf3-43d4-b9fb-0b312c9c0737/operator/0.log" Jan 23 09:16:06 crc kubenswrapper[4899]: I0123 09:16:06.542244 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-vgmkw_28fbc9c0-2fdd-46d3-bc64-bcec54804e17/perses-operator/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.288447 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/util/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.544612 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/pull/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.556614 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/util/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.562329 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/pull/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.709029 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/util/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.740237 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/pull/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.754201 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akrz8t_d8537859-a9d3-426f-9f67-631b7b5763fc/extract/0.log" Jan 23 09:16:19 crc kubenswrapper[4899]: I0123 09:16:19.879961 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.043191 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.055037 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.060611 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.201969 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.209810 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.229328 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fcfdq7_ee87fff6-abaf-4b8b-ace1-d2440eb834ed/extract/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.415052 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.612553 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.612782 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.631410 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.812012 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/util/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.817462 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/pull/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.825499 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5exs6g5_327637d2-083d-4465-90a4-be769c435c9f/extract/0.log" Jan 23 09:16:20 crc kubenswrapper[4899]: I0123 09:16:20.991330 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/util/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.132725 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/util/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.178810 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/pull/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.183996 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/pull/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.329847 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/util/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.347198 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/pull/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.363530 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc2ch_6d5136e3-cafc-4ee7-99d2-9bcc9ae4f73a/extract/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.532399 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-utilities/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.679772 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-utilities/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.685592 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-content/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.715072 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-content/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.857995 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-utilities/0.log" Jan 23 09:16:21 crc kubenswrapper[4899]: I0123 09:16:21.910043 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/extract-content/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.060084 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-44llv_58968c93-e09f-4d90-acf8-4bb499beb6e5/registry-server/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.123489 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-utilities/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.282055 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-content/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.289937 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-utilities/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.306621 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-content/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.545482 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-content/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.545926 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/extract-utilities/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.747928 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-p45p8_12cc03b0-6fca-4176-9a46-157b58c959d0/marketplace-operator/0.log" Jan 23 09:16:22 crc kubenswrapper[4899]: I0123 09:16:22.881951 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-utilities/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.008895 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q5nf4_683effa2-db47-4531-9aad-da9b77a8daf9/registry-server/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.032121 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-content/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.069850 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-utilities/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.079019 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-content/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.264657 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-content/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.287725 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/extract-utilities/0.log" Jan 23 09:16:23 crc kubenswrapper[4899]: I0123 09:16:23.653828 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fq4bf_cf74343d-e2ff-46ab-82d4-fe1fc1cc8b56/registry-server/0.log" Jan 23 09:16:34 crc kubenswrapper[4899]: I0123 09:16:34.377758 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-5dpg7_a9bf5caa-85e0-4b70-8145-d6c25fa15999/prometheus-operator/0.log" Jan 23 09:16:34 crc kubenswrapper[4899]: I0123 09:16:34.401035 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5dbb8d4666-xxpzw_e90996a1-43d9-44c4-b23e-e8b110764326/prometheus-operator-admission-webhook/0.log" Jan 23 09:16:34 crc kubenswrapper[4899]: I0123 09:16:34.445132 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5dbb8d4666-s44wd_ab9f7afb-bc13-43ce-b218-bdb18a103b51/prometheus-operator-admission-webhook/0.log" Jan 23 09:16:34 crc kubenswrapper[4899]: I0123 09:16:34.546619 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-vgmkw_28fbc9c0-2fdd-46d3-bc64-bcec54804e17/perses-operator/0.log" Jan 23 09:16:34 crc kubenswrapper[4899]: I0123 09:16:34.552296 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-g9rtj_6a152f52-cbf3-43d4-b9fb-0b312c9c0737/operator/0.log" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.218305 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:37 crc kubenswrapper[4899]: E0123 09:16:37.219513 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ed0ab1-d385-4204-85a9-724f2b327425" containerName="collect-profiles" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219535 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ed0ab1-d385-4204-85a9-724f2b327425" containerName="collect-profiles" Jan 23 09:16:37 crc kubenswrapper[4899]: E0123 09:16:37.219551 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="registry-server" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219558 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="registry-server" Jan 23 09:16:37 crc kubenswrapper[4899]: E0123 09:16:37.219572 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="extract-content" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219579 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="extract-content" Jan 23 09:16:37 crc kubenswrapper[4899]: E0123 09:16:37.219597 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="extract-utilities" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219606 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="extract-utilities" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219738 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="56ce0e86-a8dd-4076-b39a-3e62ad092ac9" containerName="registry-server" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.219760 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ed0ab1-d385-4204-85a9-724f2b327425" containerName="collect-profiles" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.220841 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.232108 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.373192 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.373530 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgvzd\" (UniqueName: \"kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.373590 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.475067 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.475138 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgvzd\" (UniqueName: \"kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.475193 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.475744 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.475762 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.498305 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgvzd\" (UniqueName: \"kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd\") pod \"certified-operators-zlmqh\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:37 crc kubenswrapper[4899]: I0123 09:16:37.562484 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:38 crc kubenswrapper[4899]: I0123 09:16:38.019606 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:38 crc kubenswrapper[4899]: I0123 09:16:38.930828 4899 generic.go:334] "Generic (PLEG): container finished" podID="95af017d-2f18-4f60-a746-f202a19653e8" containerID="33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e" exitCode=0 Jan 23 09:16:38 crc kubenswrapper[4899]: I0123 09:16:38.931051 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerDied","Data":"33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e"} Jan 23 09:16:38 crc kubenswrapper[4899]: I0123 09:16:38.931075 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerStarted","Data":"38e142cfd59b1db6c7abd9a5f898e20f7ec76d8f7ea0aed8ef476b202f339546"} Jan 23 09:16:38 crc kubenswrapper[4899]: I0123 09:16:38.932870 4899 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 23 09:16:39 crc kubenswrapper[4899]: I0123 09:16:39.941097 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerStarted","Data":"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a"} Jan 23 09:16:40 crc kubenswrapper[4899]: I0123 09:16:40.948396 4899 generic.go:334] "Generic (PLEG): container finished" podID="95af017d-2f18-4f60-a746-f202a19653e8" containerID="4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a" exitCode=0 Jan 23 09:16:40 crc kubenswrapper[4899]: I0123 09:16:40.948433 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerDied","Data":"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a"} Jan 23 09:16:41 crc kubenswrapper[4899]: I0123 09:16:41.957264 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerStarted","Data":"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99"} Jan 23 09:16:41 crc kubenswrapper[4899]: I0123 09:16:41.974479 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zlmqh" podStartSLOduration=2.3972758069999998 podStartE2EDuration="4.974457575s" podCreationTimestamp="2026-01-23 09:16:37 +0000 UTC" firstStartedPulling="2026-01-23 09:16:38.932665137 +0000 UTC m=+1485.799635524" lastFinishedPulling="2026-01-23 09:16:41.509846875 +0000 UTC m=+1488.376817292" observedRunningTime="2026-01-23 09:16:41.973830408 +0000 UTC m=+1488.840800785" watchObservedRunningTime="2026-01-23 09:16:41.974457575 +0000 UTC m=+1488.841427972" Jan 23 09:16:47 crc kubenswrapper[4899]: I0123 09:16:47.562855 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:47 crc kubenswrapper[4899]: I0123 09:16:47.566166 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:47 crc kubenswrapper[4899]: I0123 09:16:47.612885 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:48 crc kubenswrapper[4899]: I0123 09:16:48.062404 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:48 crc kubenswrapper[4899]: I0123 09:16:48.107802 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:50 crc kubenswrapper[4899]: I0123 09:16:50.013450 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zlmqh" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="registry-server" containerID="cri-o://273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99" gracePeriod=2 Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.464515 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.599909 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content\") pod \"95af017d-2f18-4f60-a746-f202a19653e8\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.600320 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities\") pod \"95af017d-2f18-4f60-a746-f202a19653e8\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.600481 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgvzd\" (UniqueName: \"kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd\") pod \"95af017d-2f18-4f60-a746-f202a19653e8\" (UID: \"95af017d-2f18-4f60-a746-f202a19653e8\") " Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.601329 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities" (OuterVolumeSpecName: "utilities") pod "95af017d-2f18-4f60-a746-f202a19653e8" (UID: "95af017d-2f18-4f60-a746-f202a19653e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.605792 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd" (OuterVolumeSpecName: "kube-api-access-zgvzd") pod "95af017d-2f18-4f60-a746-f202a19653e8" (UID: "95af017d-2f18-4f60-a746-f202a19653e8"). InnerVolumeSpecName "kube-api-access-zgvzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.648087 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95af017d-2f18-4f60-a746-f202a19653e8" (UID: "95af017d-2f18-4f60-a746-f202a19653e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.702273 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.702595 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95af017d-2f18-4f60-a746-f202a19653e8-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:16:51 crc kubenswrapper[4899]: I0123 09:16:51.702742 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgvzd\" (UniqueName: \"kubernetes.io/projected/95af017d-2f18-4f60-a746-f202a19653e8-kube-api-access-zgvzd\") on node \"crc\" DevicePath \"\"" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.028624 4899 generic.go:334] "Generic (PLEG): container finished" podID="95af017d-2f18-4f60-a746-f202a19653e8" containerID="273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99" exitCode=0 Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.029248 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerDied","Data":"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99"} Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.029323 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlmqh" event={"ID":"95af017d-2f18-4f60-a746-f202a19653e8","Type":"ContainerDied","Data":"38e142cfd59b1db6c7abd9a5f898e20f7ec76d8f7ea0aed8ef476b202f339546"} Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.029347 4899 scope.go:117] "RemoveContainer" containerID="273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.029251 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlmqh" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.060752 4899 scope.go:117] "RemoveContainer" containerID="4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.070279 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.076055 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zlmqh"] Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.086979 4899 scope.go:117] "RemoveContainer" containerID="33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.109025 4899 scope.go:117] "RemoveContainer" containerID="273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99" Jan 23 09:16:52 crc kubenswrapper[4899]: E0123 09:16:52.109540 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99\": container with ID starting with 273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99 not found: ID does not exist" containerID="273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.109573 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99"} err="failed to get container status \"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99\": rpc error: code = NotFound desc = could not find container \"273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99\": container with ID starting with 273daae0effd603c158ed320d2faa420d07e2bfd164f262b413f3104fddb9e99 not found: ID does not exist" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.109593 4899 scope.go:117] "RemoveContainer" containerID="4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a" Jan 23 09:16:52 crc kubenswrapper[4899]: E0123 09:16:52.110082 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a\": container with ID starting with 4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a not found: ID does not exist" containerID="4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.110191 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a"} err="failed to get container status \"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a\": rpc error: code = NotFound desc = could not find container \"4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a\": container with ID starting with 4235a172ccb3d138edb81e6fb08346ae5f0cdcc0ae061eb245349c6dc87e592a not found: ID does not exist" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.110566 4899 scope.go:117] "RemoveContainer" containerID="33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e" Jan 23 09:16:52 crc kubenswrapper[4899]: E0123 09:16:52.110996 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e\": container with ID starting with 33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e not found: ID does not exist" containerID="33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e" Jan 23 09:16:52 crc kubenswrapper[4899]: I0123 09:16:52.111049 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e"} err="failed to get container status \"33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e\": rpc error: code = NotFound desc = could not find container \"33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e\": container with ID starting with 33604ff82b53057007f649843493739f762ea70070f4d69a8b8f47090a1b355e not found: ID does not exist" Jan 23 09:16:53 crc kubenswrapper[4899]: I0123 09:16:53.531564 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95af017d-2f18-4f60-a746-f202a19653e8" path="/var/lib/kubelet/pods/95af017d-2f18-4f60-a746-f202a19653e8/volumes" Jan 23 09:17:28 crc kubenswrapper[4899]: I0123 09:17:28.306493 4899 generic.go:334] "Generic (PLEG): container finished" podID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerID="10ab614b799ec1ff38e96f74c6c7c2e7f4fdbe50265ebc57e68921c4441c310e" exitCode=0 Jan 23 09:17:28 crc kubenswrapper[4899]: I0123 09:17:28.306576 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qsppb/must-gather-m9dh9" event={"ID":"bfbde8db-491d-4d66-849b-c0ad6abcc433","Type":"ContainerDied","Data":"10ab614b799ec1ff38e96f74c6c7c2e7f4fdbe50265ebc57e68921c4441c310e"} Jan 23 09:17:28 crc kubenswrapper[4899]: I0123 09:17:28.308579 4899 scope.go:117] "RemoveContainer" containerID="10ab614b799ec1ff38e96f74c6c7c2e7f4fdbe50265ebc57e68921c4441c310e" Jan 23 09:17:28 crc kubenswrapper[4899]: I0123 09:17:28.862202 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qsppb_must-gather-m9dh9_bfbde8db-491d-4d66-849b-c0ad6abcc433/gather/0.log" Jan 23 09:17:31 crc kubenswrapper[4899]: I0123 09:17:31.686392 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:17:31 crc kubenswrapper[4899]: I0123 09:17:31.686753 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:17:35 crc kubenswrapper[4899]: I0123 09:17:35.974487 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qsppb/must-gather-m9dh9"] Jan 23 09:17:35 crc kubenswrapper[4899]: I0123 09:17:35.975383 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qsppb/must-gather-m9dh9" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="copy" containerID="cri-o://fb4c0a30ee9849139c728da10db1d8a0e813df0438bf9f4155afb80aed2d0f19" gracePeriod=2 Jan 23 09:17:35 crc kubenswrapper[4899]: I0123 09:17:35.980361 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qsppb/must-gather-m9dh9"] Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.370506 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qsppb_must-gather-m9dh9_bfbde8db-491d-4d66-849b-c0ad6abcc433/copy/0.log" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.371329 4899 generic.go:334] "Generic (PLEG): container finished" podID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerID="fb4c0a30ee9849139c728da10db1d8a0e813df0438bf9f4155afb80aed2d0f19" exitCode=143 Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.438854 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qsppb_must-gather-m9dh9_bfbde8db-491d-4d66-849b-c0ad6abcc433/copy/0.log" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.439294 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.568778 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj8kq\" (UniqueName: \"kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq\") pod \"bfbde8db-491d-4d66-849b-c0ad6abcc433\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.569075 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output\") pod \"bfbde8db-491d-4d66-849b-c0ad6abcc433\" (UID: \"bfbde8db-491d-4d66-849b-c0ad6abcc433\") " Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.578422 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq" (OuterVolumeSpecName: "kube-api-access-cj8kq") pod "bfbde8db-491d-4d66-849b-c0ad6abcc433" (UID: "bfbde8db-491d-4d66-849b-c0ad6abcc433"). InnerVolumeSpecName "kube-api-access-cj8kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.639692 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bfbde8db-491d-4d66-849b-c0ad6abcc433" (UID: "bfbde8db-491d-4d66-849b-c0ad6abcc433"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.672364 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj8kq\" (UniqueName: \"kubernetes.io/projected/bfbde8db-491d-4d66-849b-c0ad6abcc433-kube-api-access-cj8kq\") on node \"crc\" DevicePath \"\"" Jan 23 09:17:36 crc kubenswrapper[4899]: I0123 09:17:36.672407 4899 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bfbde8db-491d-4d66-849b-c0ad6abcc433-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 23 09:17:37 crc kubenswrapper[4899]: I0123 09:17:37.382709 4899 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qsppb_must-gather-m9dh9_bfbde8db-491d-4d66-849b-c0ad6abcc433/copy/0.log" Jan 23 09:17:37 crc kubenswrapper[4899]: I0123 09:17:37.383688 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qsppb/must-gather-m9dh9" Jan 23 09:17:37 crc kubenswrapper[4899]: I0123 09:17:37.383716 4899 scope.go:117] "RemoveContainer" containerID="fb4c0a30ee9849139c728da10db1d8a0e813df0438bf9f4155afb80aed2d0f19" Jan 23 09:17:37 crc kubenswrapper[4899]: I0123 09:17:37.406301 4899 scope.go:117] "RemoveContainer" containerID="10ab614b799ec1ff38e96f74c6c7c2e7f4fdbe50265ebc57e68921c4441c310e" Jan 23 09:17:37 crc kubenswrapper[4899]: I0123 09:17:37.525870 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" path="/var/lib/kubelet/pods/bfbde8db-491d-4d66-849b-c0ad6abcc433/volumes" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.859217 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:17:56 crc kubenswrapper[4899]: E0123 09:17:56.859977 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="gather" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.859989 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="gather" Jan 23 09:17:56 crc kubenswrapper[4899]: E0123 09:17:56.860001 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="extract-utilities" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860008 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="extract-utilities" Jan 23 09:17:56 crc kubenswrapper[4899]: E0123 09:17:56.860027 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="extract-content" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860033 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="extract-content" Jan 23 09:17:56 crc kubenswrapper[4899]: E0123 09:17:56.860041 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="copy" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860046 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="copy" Jan 23 09:17:56 crc kubenswrapper[4899]: E0123 09:17:56.860059 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="registry-server" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860064 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="registry-server" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860188 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="95af017d-2f18-4f60-a746-f202a19653e8" containerName="registry-server" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860201 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="gather" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.860286 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfbde8db-491d-4d66-849b-c0ad6abcc433" containerName="copy" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.861130 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.872832 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.992195 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.992433 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:56 crc kubenswrapper[4899]: I0123 09:17:56.992558 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmxfb\" (UniqueName: \"kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.094166 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.094261 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.094307 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmxfb\" (UniqueName: \"kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.094716 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.094945 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.119571 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmxfb\" (UniqueName: \"kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb\") pod \"community-operators-pc8n9\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.181531 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:17:57 crc kubenswrapper[4899]: I0123 09:17:57.661785 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:17:58 crc kubenswrapper[4899]: I0123 09:17:58.558862 4899 generic.go:334] "Generic (PLEG): container finished" podID="0ad08b89-e369-469d-ac87-5c85ef220444" containerID="db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc" exitCode=0 Jan 23 09:17:58 crc kubenswrapper[4899]: I0123 09:17:58.558935 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerDied","Data":"db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc"} Jan 23 09:17:58 crc kubenswrapper[4899]: I0123 09:17:58.559296 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerStarted","Data":"0aa685cfc44d5ca4dfdf0d3849c25c0ce3a94599fe80b45ad528eabb010fb77a"} Jan 23 09:17:59 crc kubenswrapper[4899]: I0123 09:17:59.567703 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerStarted","Data":"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d"} Jan 23 09:18:00 crc kubenswrapper[4899]: I0123 09:18:00.577313 4899 generic.go:334] "Generic (PLEG): container finished" podID="0ad08b89-e369-469d-ac87-5c85ef220444" containerID="91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d" exitCode=0 Jan 23 09:18:00 crc kubenswrapper[4899]: I0123 09:18:00.577367 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerDied","Data":"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d"} Jan 23 09:18:01 crc kubenswrapper[4899]: I0123 09:18:01.585791 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerStarted","Data":"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac"} Jan 23 09:18:01 crc kubenswrapper[4899]: I0123 09:18:01.603141 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pc8n9" podStartSLOduration=3.0110775 podStartE2EDuration="5.60312486s" podCreationTimestamp="2026-01-23 09:17:56 +0000 UTC" firstStartedPulling="2026-01-23 09:17:58.562771711 +0000 UTC m=+1565.429742098" lastFinishedPulling="2026-01-23 09:18:01.154819061 +0000 UTC m=+1568.021789458" observedRunningTime="2026-01-23 09:18:01.601013632 +0000 UTC m=+1568.467984029" watchObservedRunningTime="2026-01-23 09:18:01.60312486 +0000 UTC m=+1568.470095247" Jan 23 09:18:01 crc kubenswrapper[4899]: I0123 09:18:01.686189 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:18:01 crc kubenswrapper[4899]: I0123 09:18:01.686289 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:18:07 crc kubenswrapper[4899]: I0123 09:18:07.181765 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:07 crc kubenswrapper[4899]: I0123 09:18:07.182303 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:07 crc kubenswrapper[4899]: I0123 09:18:07.221700 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:07 crc kubenswrapper[4899]: I0123 09:18:07.667767 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:07 crc kubenswrapper[4899]: I0123 09:18:07.732605 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:18:09 crc kubenswrapper[4899]: I0123 09:18:09.634344 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pc8n9" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="registry-server" containerID="cri-o://c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac" gracePeriod=2 Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.132178 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.309309 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmxfb\" (UniqueName: \"kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb\") pod \"0ad08b89-e369-469d-ac87-5c85ef220444\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.309412 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content\") pod \"0ad08b89-e369-469d-ac87-5c85ef220444\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.309545 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities\") pod \"0ad08b89-e369-469d-ac87-5c85ef220444\" (UID: \"0ad08b89-e369-469d-ac87-5c85ef220444\") " Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.311331 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities" (OuterVolumeSpecName: "utilities") pod "0ad08b89-e369-469d-ac87-5c85ef220444" (UID: "0ad08b89-e369-469d-ac87-5c85ef220444"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.315261 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb" (OuterVolumeSpecName: "kube-api-access-kmxfb") pod "0ad08b89-e369-469d-ac87-5c85ef220444" (UID: "0ad08b89-e369-469d-ac87-5c85ef220444"). InnerVolumeSpecName "kube-api-access-kmxfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.369998 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ad08b89-e369-469d-ac87-5c85ef220444" (UID: "0ad08b89-e369-469d-ac87-5c85ef220444"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.411153 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmxfb\" (UniqueName: \"kubernetes.io/projected/0ad08b89-e369-469d-ac87-5c85ef220444-kube-api-access-kmxfb\") on node \"crc\" DevicePath \"\"" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.411191 4899 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.411202 4899 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad08b89-e369-469d-ac87-5c85ef220444-utilities\") on node \"crc\" DevicePath \"\"" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.642757 4899 generic.go:334] "Generic (PLEG): container finished" podID="0ad08b89-e369-469d-ac87-5c85ef220444" containerID="c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac" exitCode=0 Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.642819 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerDied","Data":"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac"} Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.642848 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc8n9" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.642869 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc8n9" event={"ID":"0ad08b89-e369-469d-ac87-5c85ef220444","Type":"ContainerDied","Data":"0aa685cfc44d5ca4dfdf0d3849c25c0ce3a94599fe80b45ad528eabb010fb77a"} Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.642889 4899 scope.go:117] "RemoveContainer" containerID="c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.678035 4899 scope.go:117] "RemoveContainer" containerID="91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.691261 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.696994 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pc8n9"] Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.723232 4899 scope.go:117] "RemoveContainer" containerID="db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.739988 4899 scope.go:117] "RemoveContainer" containerID="c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac" Jan 23 09:18:10 crc kubenswrapper[4899]: E0123 09:18:10.740474 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac\": container with ID starting with c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac not found: ID does not exist" containerID="c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.740526 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac"} err="failed to get container status \"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac\": rpc error: code = NotFound desc = could not find container \"c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac\": container with ID starting with c68070602cfb908edc2fe889465cf75b2b90b8c7ad1df7974412cb2f03edcfac not found: ID does not exist" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.740559 4899 scope.go:117] "RemoveContainer" containerID="91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d" Jan 23 09:18:10 crc kubenswrapper[4899]: E0123 09:18:10.741022 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d\": container with ID starting with 91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d not found: ID does not exist" containerID="91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.741049 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d"} err="failed to get container status \"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d\": rpc error: code = NotFound desc = could not find container \"91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d\": container with ID starting with 91ef9eb70ec6ad2df4d53e62cc6b2b7bfe1489fcf5fe7e66a4521bf21708941d not found: ID does not exist" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.741067 4899 scope.go:117] "RemoveContainer" containerID="db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc" Jan 23 09:18:10 crc kubenswrapper[4899]: E0123 09:18:10.741452 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc\": container with ID starting with db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc not found: ID does not exist" containerID="db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc" Jan 23 09:18:10 crc kubenswrapper[4899]: I0123 09:18:10.741506 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc"} err="failed to get container status \"db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc\": rpc error: code = NotFound desc = could not find container \"db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc\": container with ID starting with db4e673b7953e6657eec5eb262f4ce2f3b6d14a5f411891cabfe2a7799135fbc not found: ID does not exist" Jan 23 09:18:11 crc kubenswrapper[4899]: I0123 09:18:11.522121 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" path="/var/lib/kubelet/pods/0ad08b89-e369-469d-ac87-5c85ef220444/volumes" Jan 23 09:18:31 crc kubenswrapper[4899]: I0123 09:18:31.686243 4899 patch_prober.go:28] interesting pod/machine-config-daemon-x6wwr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 23 09:18:31 crc kubenswrapper[4899]: I0123 09:18:31.686746 4899 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 23 09:18:31 crc kubenswrapper[4899]: I0123 09:18:31.686791 4899 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" Jan 23 09:18:31 crc kubenswrapper[4899]: I0123 09:18:31.687519 4899 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3"} pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 23 09:18:31 crc kubenswrapper[4899]: I0123 09:18:31.687587 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerName="machine-config-daemon" containerID="cri-o://49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" gracePeriod=600 Jan 23 09:18:31 crc kubenswrapper[4899]: E0123 09:18:31.818669 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:18:32 crc kubenswrapper[4899]: I0123 09:18:32.799007 4899 generic.go:334] "Generic (PLEG): container finished" podID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" exitCode=0 Jan 23 09:18:32 crc kubenswrapper[4899]: I0123 09:18:32.799072 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" event={"ID":"57f7d8bc-8a1d-4b44-be6e-d702c34203c8","Type":"ContainerDied","Data":"49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3"} Jan 23 09:18:32 crc kubenswrapper[4899]: I0123 09:18:32.799132 4899 scope.go:117] "RemoveContainer" containerID="15e3a3d08999627f66e6ecc42a3229f6a5454ad964ad7771dd31fedbf3810ebc" Jan 23 09:18:32 crc kubenswrapper[4899]: I0123 09:18:32.799734 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:18:32 crc kubenswrapper[4899]: E0123 09:18:32.800255 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:18:43 crc kubenswrapper[4899]: I0123 09:18:43.519135 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:18:43 crc kubenswrapper[4899]: E0123 09:18:43.519908 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:18:58 crc kubenswrapper[4899]: I0123 09:18:58.514305 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:18:58 crc kubenswrapper[4899]: E0123 09:18:58.515004 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:19:13 crc kubenswrapper[4899]: I0123 09:19:13.523691 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:19:13 crc kubenswrapper[4899]: E0123 09:19:13.524539 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:19:26 crc kubenswrapper[4899]: I0123 09:19:26.515351 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:19:26 crc kubenswrapper[4899]: E0123 09:19:26.516151 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.287435 4899 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:33 crc kubenswrapper[4899]: E0123 09:19:33.288165 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="registry-server" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.288179 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="registry-server" Jan 23 09:19:33 crc kubenswrapper[4899]: E0123 09:19:33.288195 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="extract-utilities" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.288201 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="extract-utilities" Jan 23 09:19:33 crc kubenswrapper[4899]: E0123 09:19:33.288233 4899 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="extract-content" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.288239 4899 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="extract-content" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.288343 4899 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ad08b89-e369-469d-ac87-5c85ef220444" containerName="registry-server" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.288799 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.298020 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.490579 4899 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrx8\" (UniqueName: \"kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8\") pod \"infrawatch-operators-85csh\" (UID: \"48541030-38c4-43f5-8bfc-a49efa3a1627\") " pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.592315 4899 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrx8\" (UniqueName: \"kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8\") pod \"infrawatch-operators-85csh\" (UID: \"48541030-38c4-43f5-8bfc-a49efa3a1627\") " pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.615808 4899 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrx8\" (UniqueName: \"kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8\") pod \"infrawatch-operators-85csh\" (UID: \"48541030-38c4-43f5-8bfc-a49efa3a1627\") " pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:33 crc kubenswrapper[4899]: I0123 09:19:33.915199 4899 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:34 crc kubenswrapper[4899]: I0123 09:19:34.171370 4899 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:34 crc kubenswrapper[4899]: I0123 09:19:34.244849 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-85csh" event={"ID":"48541030-38c4-43f5-8bfc-a49efa3a1627","Type":"ContainerStarted","Data":"1660986d2e6e8d8bacee1b93713aa1e3dba14bbccaef9bc6d29f4813fc61b85b"} Jan 23 09:19:35 crc kubenswrapper[4899]: I0123 09:19:35.254379 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-85csh" event={"ID":"48541030-38c4-43f5-8bfc-a49efa3a1627","Type":"ContainerStarted","Data":"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980"} Jan 23 09:19:40 crc kubenswrapper[4899]: I0123 09:19:40.515296 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:19:40 crc kubenswrapper[4899]: E0123 09:19:40.515855 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:19:43 crc kubenswrapper[4899]: I0123 09:19:43.916544 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:43 crc kubenswrapper[4899]: I0123 09:19:43.916708 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:43 crc kubenswrapper[4899]: I0123 09:19:43.946379 4899 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:43 crc kubenswrapper[4899]: I0123 09:19:43.967164 4899 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-85csh" podStartSLOduration=10.854554228 podStartE2EDuration="10.967143044s" podCreationTimestamp="2026-01-23 09:19:33 +0000 UTC" firstStartedPulling="2026-01-23 09:19:34.184814389 +0000 UTC m=+1661.051784776" lastFinishedPulling="2026-01-23 09:19:34.297403205 +0000 UTC m=+1661.164373592" observedRunningTime="2026-01-23 09:19:35.269806891 +0000 UTC m=+1662.136777278" watchObservedRunningTime="2026-01-23 09:19:43.967143044 +0000 UTC m=+1670.834113441" Jan 23 09:19:44 crc kubenswrapper[4899]: I0123 09:19:44.339683 4899 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:44 crc kubenswrapper[4899]: I0123 09:19:44.382377 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:46 crc kubenswrapper[4899]: I0123 09:19:46.328378 4899 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-85csh" podUID="48541030-38c4-43f5-8bfc-a49efa3a1627" containerName="registry-server" containerID="cri-o://a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980" gracePeriod=2 Jan 23 09:19:46 crc kubenswrapper[4899]: I0123 09:19:46.676949 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:46 crc kubenswrapper[4899]: I0123 09:19:46.793485 4899 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxrx8\" (UniqueName: \"kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8\") pod \"48541030-38c4-43f5-8bfc-a49efa3a1627\" (UID: \"48541030-38c4-43f5-8bfc-a49efa3a1627\") " Jan 23 09:19:46 crc kubenswrapper[4899]: I0123 09:19:46.798841 4899 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8" (OuterVolumeSpecName: "kube-api-access-gxrx8") pod "48541030-38c4-43f5-8bfc-a49efa3a1627" (UID: "48541030-38c4-43f5-8bfc-a49efa3a1627"). InnerVolumeSpecName "kube-api-access-gxrx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 23 09:19:46 crc kubenswrapper[4899]: I0123 09:19:46.895229 4899 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxrx8\" (UniqueName: \"kubernetes.io/projected/48541030-38c4-43f5-8bfc-a49efa3a1627-kube-api-access-gxrx8\") on node \"crc\" DevicePath \"\"" Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.336052 4899 generic.go:334] "Generic (PLEG): container finished" podID="48541030-38c4-43f5-8bfc-a49efa3a1627" containerID="a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980" exitCode=0 Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.336150 4899 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-85csh" Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.336152 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-85csh" event={"ID":"48541030-38c4-43f5-8bfc-a49efa3a1627","Type":"ContainerDied","Data":"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980"} Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.337451 4899 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-85csh" event={"ID":"48541030-38c4-43f5-8bfc-a49efa3a1627","Type":"ContainerDied","Data":"1660986d2e6e8d8bacee1b93713aa1e3dba14bbccaef9bc6d29f4813fc61b85b"} Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.337487 4899 scope.go:117] "RemoveContainer" containerID="a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980" Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.354873 4899 scope.go:117] "RemoveContainer" containerID="a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980" Jan 23 09:19:47 crc kubenswrapper[4899]: E0123 09:19:47.355576 4899 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980\": container with ID starting with a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980 not found: ID does not exist" containerID="a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980" Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.355669 4899 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980"} err="failed to get container status \"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980\": rpc error: code = NotFound desc = could not find container \"a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980\": container with ID starting with a248bb3d9457dcd9bb11d88112e877f84a3698bd5c6ecafbc28422955fa8c980 not found: ID does not exist" Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.382424 4899 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.406807 4899 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-85csh"] Jan 23 09:19:47 crc kubenswrapper[4899]: I0123 09:19:47.523022 4899 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48541030-38c4-43f5-8bfc-a49efa3a1627" path="/var/lib/kubelet/pods/48541030-38c4-43f5-8bfc-a49efa3a1627/volumes" Jan 23 09:19:52 crc kubenswrapper[4899]: I0123 09:19:52.514939 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:19:52 crc kubenswrapper[4899]: E0123 09:19:52.515431 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:20:04 crc kubenswrapper[4899]: I0123 09:20:04.514940 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:20:04 crc kubenswrapper[4899]: E0123 09:20:04.515623 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" Jan 23 09:20:16 crc kubenswrapper[4899]: I0123 09:20:16.515094 4899 scope.go:117] "RemoveContainer" containerID="49e20a3255806efbaa5cc0c93ab8216b0bfe4520d0136f7892b609101ac56ff3" Jan 23 09:20:16 crc kubenswrapper[4899]: E0123 09:20:16.515832 4899 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6wwr_openshift-machine-config-operator(57f7d8bc-8a1d-4b44-be6e-d702c34203c8)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6wwr" podUID="57f7d8bc-8a1d-4b44-be6e-d702c34203c8" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134636547024463 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134636547017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134632744016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134632744015466 5ustar corecore